var/home/core/zuul-output/0000755000175000017500000000000015113367050014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113400745015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005372451115113400736017705 0ustar rootrootDec 01 19:39:23 crc systemd[1]: Starting Kubernetes Kubelet... Dec 01 19:39:23 crc restorecon[4683]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:23 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 01 19:39:24 crc restorecon[4683]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 01 19:39:25 crc kubenswrapper[4960]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.100812 4960 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108291 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108367 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108380 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108391 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108408 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108420 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108433 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108446 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108457 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108468 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108479 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108490 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108500 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108510 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108520 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108529 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108545 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108559 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108569 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108581 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108592 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108602 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108612 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108622 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108634 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108647 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108659 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108673 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108684 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108705 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108722 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108734 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108747 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108758 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108768 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108778 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108788 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108803 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108820 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108833 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108844 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108855 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108867 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108878 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108890 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108900 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108911 4960 feature_gate.go:330] unrecognized feature gate: Example Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108920 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108931 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108941 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108951 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108961 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108971 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108982 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.108992 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109002 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109014 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109029 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109039 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109049 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109060 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109071 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109084 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109094 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109106 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109155 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109166 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109178 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109188 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109199 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.109213 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109485 4960 flags.go:64] FLAG: --address="0.0.0.0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109517 4960 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109543 4960 flags.go:64] FLAG: --anonymous-auth="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109560 4960 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109576 4960 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109589 4960 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109606 4960 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109622 4960 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109636 4960 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109650 4960 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109664 4960 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109681 4960 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109694 4960 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109706 4960 flags.go:64] FLAG: --cgroup-root="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109718 4960 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109729 4960 flags.go:64] FLAG: --client-ca-file="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109741 4960 flags.go:64] FLAG: --cloud-config="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109756 4960 flags.go:64] FLAG: --cloud-provider="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109769 4960 flags.go:64] FLAG: --cluster-dns="[]" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109786 4960 flags.go:64] FLAG: --cluster-domain="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109798 4960 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109810 4960 flags.go:64] FLAG: --config-dir="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109822 4960 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109835 4960 flags.go:64] FLAG: --container-log-max-files="5" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109854 4960 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109865 4960 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109878 4960 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109890 4960 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109902 4960 flags.go:64] FLAG: --contention-profiling="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109914 4960 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109926 4960 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109938 4960 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109950 4960 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109965 4960 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109979 4960 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.109991 4960 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110002 4960 flags.go:64] FLAG: --enable-load-reader="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110014 4960 flags.go:64] FLAG: --enable-server="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110025 4960 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110040 4960 flags.go:64] FLAG: --event-burst="100" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110051 4960 flags.go:64] FLAG: --event-qps="50" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110060 4960 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110070 4960 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110079 4960 flags.go:64] FLAG: --eviction-hard="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110090 4960 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110100 4960 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110138 4960 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110151 4960 flags.go:64] FLAG: --eviction-soft="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110160 4960 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110169 4960 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110179 4960 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110188 4960 flags.go:64] FLAG: --experimental-mounter-path="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110197 4960 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110206 4960 flags.go:64] FLAG: --fail-swap-on="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110215 4960 flags.go:64] FLAG: --feature-gates="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110227 4960 flags.go:64] FLAG: --file-check-frequency="20s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110237 4960 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110247 4960 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110256 4960 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110266 4960 flags.go:64] FLAG: --healthz-port="10248" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110276 4960 flags.go:64] FLAG: --help="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110285 4960 flags.go:64] FLAG: --hostname-override="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110294 4960 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110304 4960 flags.go:64] FLAG: --http-check-frequency="20s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110314 4960 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110324 4960 flags.go:64] FLAG: --image-credential-provider-config="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110333 4960 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110343 4960 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110352 4960 flags.go:64] FLAG: --image-service-endpoint="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110361 4960 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110371 4960 flags.go:64] FLAG: --kube-api-burst="100" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110381 4960 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110391 4960 flags.go:64] FLAG: --kube-api-qps="50" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110401 4960 flags.go:64] FLAG: --kube-reserved="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110410 4960 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110419 4960 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110429 4960 flags.go:64] FLAG: --kubelet-cgroups="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110438 4960 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110448 4960 flags.go:64] FLAG: --lock-file="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110457 4960 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110466 4960 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110476 4960 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110492 4960 flags.go:64] FLAG: --log-json-split-stream="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110503 4960 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110512 4960 flags.go:64] FLAG: --log-text-split-stream="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110521 4960 flags.go:64] FLAG: --logging-format="text" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110530 4960 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110540 4960 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110549 4960 flags.go:64] FLAG: --manifest-url="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110559 4960 flags.go:64] FLAG: --manifest-url-header="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110573 4960 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110583 4960 flags.go:64] FLAG: --max-open-files="1000000" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110596 4960 flags.go:64] FLAG: --max-pods="110" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110606 4960 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110616 4960 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110626 4960 flags.go:64] FLAG: --memory-manager-policy="None" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110635 4960 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110645 4960 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110655 4960 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110664 4960 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110693 4960 flags.go:64] FLAG: --node-status-max-images="50" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110702 4960 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110712 4960 flags.go:64] FLAG: --oom-score-adj="-999" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110722 4960 flags.go:64] FLAG: --pod-cidr="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110730 4960 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110744 4960 flags.go:64] FLAG: --pod-manifest-path="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110753 4960 flags.go:64] FLAG: --pod-max-pids="-1" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110763 4960 flags.go:64] FLAG: --pods-per-core="0" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110772 4960 flags.go:64] FLAG: --port="10250" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110781 4960 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110790 4960 flags.go:64] FLAG: --provider-id="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110799 4960 flags.go:64] FLAG: --qos-reserved="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110809 4960 flags.go:64] FLAG: --read-only-port="10255" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110818 4960 flags.go:64] FLAG: --register-node="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110827 4960 flags.go:64] FLAG: --register-schedulable="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110837 4960 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110854 4960 flags.go:64] FLAG: --registry-burst="10" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110864 4960 flags.go:64] FLAG: --registry-qps="5" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110873 4960 flags.go:64] FLAG: --reserved-cpus="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110883 4960 flags.go:64] FLAG: --reserved-memory="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110895 4960 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110904 4960 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110914 4960 flags.go:64] FLAG: --rotate-certificates="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110923 4960 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110933 4960 flags.go:64] FLAG: --runonce="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110941 4960 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110951 4960 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110961 4960 flags.go:64] FLAG: --seccomp-default="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110970 4960 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110980 4960 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110990 4960 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.110999 4960 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111009 4960 flags.go:64] FLAG: --storage-driver-password="root" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111018 4960 flags.go:64] FLAG: --storage-driver-secure="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111027 4960 flags.go:64] FLAG: --storage-driver-table="stats" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111036 4960 flags.go:64] FLAG: --storage-driver-user="root" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111045 4960 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111055 4960 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111065 4960 flags.go:64] FLAG: --system-cgroups="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111074 4960 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111089 4960 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111097 4960 flags.go:64] FLAG: --tls-cert-file="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111106 4960 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111141 4960 flags.go:64] FLAG: --tls-min-version="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111150 4960 flags.go:64] FLAG: --tls-private-key-file="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111159 4960 flags.go:64] FLAG: --topology-manager-policy="none" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111168 4960 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111178 4960 flags.go:64] FLAG: --topology-manager-scope="container" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111188 4960 flags.go:64] FLAG: --v="2" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111201 4960 flags.go:64] FLAG: --version="false" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111214 4960 flags.go:64] FLAG: --vmodule="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111226 4960 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.111236 4960 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111466 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111477 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111489 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111500 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111509 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111518 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111527 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111539 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111549 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111559 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111569 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111577 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111586 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111594 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111602 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111610 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111618 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111626 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111634 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111642 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111650 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111657 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111666 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111674 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111682 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111707 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111720 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111733 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111745 4960 feature_gate.go:330] unrecognized feature gate: Example Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111756 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111766 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111776 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111787 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111797 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111807 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111817 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111827 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111837 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111850 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111859 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111868 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111876 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111884 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111893 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111902 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111911 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111919 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111938 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111947 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111955 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111963 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111990 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.111998 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112007 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112016 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112024 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112032 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112045 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112054 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112062 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112069 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112077 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112088 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112138 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112148 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112156 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112164 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112172 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112180 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112189 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.112197 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.112230 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.122249 4960 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.122293 4960 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122492 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122505 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122514 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122524 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122532 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122541 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122549 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122560 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122571 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122581 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122589 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122597 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122606 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122615 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122626 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122635 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122644 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122652 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122661 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122670 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122677 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122685 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122693 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122701 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122709 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122717 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122725 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122732 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122741 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122749 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122758 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122767 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122776 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122785 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122795 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122804 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122812 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122820 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122827 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122838 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122847 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122857 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122865 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122873 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122880 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122889 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122897 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122905 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122913 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122920 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122928 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122936 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122944 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122951 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122959 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122970 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122980 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122988 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.122996 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123004 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123013 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123020 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123027 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123036 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123043 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123051 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123060 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123070 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123079 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123090 4960 feature_gate.go:330] unrecognized feature gate: Example Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123102 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.123151 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123454 4960 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123474 4960 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123489 4960 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123505 4960 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123518 4960 feature_gate.go:330] unrecognized feature gate: Example Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123530 4960 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123541 4960 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123552 4960 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123564 4960 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123574 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123582 4960 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123590 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123598 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123605 4960 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123613 4960 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123622 4960 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123630 4960 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123638 4960 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123648 4960 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123658 4960 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123668 4960 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123678 4960 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123688 4960 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123696 4960 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123705 4960 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123715 4960 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123725 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123735 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123746 4960 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123758 4960 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123768 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123778 4960 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123811 4960 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123821 4960 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123837 4960 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123847 4960 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123855 4960 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123863 4960 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123871 4960 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123879 4960 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123887 4960 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123894 4960 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123902 4960 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123910 4960 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123918 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123926 4960 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123937 4960 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123949 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123958 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123966 4960 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123976 4960 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123984 4960 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.123992 4960 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124000 4960 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124008 4960 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124016 4960 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124024 4960 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124032 4960 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124041 4960 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124048 4960 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124057 4960 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124065 4960 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124073 4960 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124081 4960 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124089 4960 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124097 4960 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124135 4960 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124143 4960 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124151 4960 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124158 4960 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.124168 4960 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.124182 4960 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.124794 4960 server.go:940] "Client rotation is on, will bootstrap in background" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.131663 4960 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.131852 4960 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.132807 4960 server.go:997] "Starting client certificate rotation" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.132857 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.133389 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 17:46:31.569621948 +0000 UTC Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.133548 4960 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 646h7m6.436080157s for next certificate rotation Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.141765 4960 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.146946 4960 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.158037 4960 log.go:25] "Validated CRI v1 runtime API" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.197729 4960 log.go:25] "Validated CRI v1 image API" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.200244 4960 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.205321 4960 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-01-19-35-01-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.205378 4960 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.236228 4960 manager.go:217] Machine: {Timestamp:2025-12-01 19:39:25.233929845 +0000 UTC m=+0.521421594 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:af0d86f2-b38c-4a68-99f8-21f571d22202 BootID:2c6e64e4-31f4-48d6-b943-ffee08ff2cb1 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:88:aa:25 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:88:aa:25 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5c:04:b8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:43:ad:8f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:77:73:8d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a4:fd:db Speed:-1 Mtu:1496} {Name:eth10 MacAddress:a6:56:8c:09:06:ea Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2a:ba:c4:ff:e5:b3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.236654 4960 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.237070 4960 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.237934 4960 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.238317 4960 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.238378 4960 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.238716 4960 topology_manager.go:138] "Creating topology manager with none policy" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.238733 4960 container_manager_linux.go:303] "Creating device plugin manager" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.239054 4960 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.239146 4960 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.239629 4960 state_mem.go:36] "Initialized new in-memory state store" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.239773 4960 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.242436 4960 kubelet.go:418] "Attempting to sync node with API server" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.242490 4960 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.242545 4960 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.242572 4960 kubelet.go:324] "Adding apiserver pod source" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.242623 4960 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.246357 4960 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.247260 4960 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.248447 4960 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249448 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249512 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249535 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249554 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249581 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.249517 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249646 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249671 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249706 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249726 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249745 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.249586 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.249792 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.249873 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249805 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.249957 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.250258 4960 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.251068 4960 server.go:1280] "Started kubelet" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.251438 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.252056 4960 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.252037 4960 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.253372 4960 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 01 19:39:25 crc systemd[1]: Started Kubernetes Kubelet. Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.254234 4960 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d2eb0a70f434c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 19:39:25.251019596 +0000 UTC m=+0.538511305,LastTimestamp:2025-12-01 19:39:25.251019596 +0000 UTC m=+0.538511305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.256642 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.256722 4960 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.256824 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 10:35:34.376427342 +0000 UTC Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.256913 4960 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 302h56m9.11951933s for next certificate rotation Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.256874 4960 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.257034 4960 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.257096 4960 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.256888 4960 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.259180 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.259337 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.262361 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="200ms" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.264686 4960 factory.go:55] Registering systemd factory Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.264736 4960 factory.go:221] Registration of the systemd container factory successfully Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.265976 4960 factory.go:153] Registering CRI-O factory Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.266032 4960 factory.go:221] Registration of the crio container factory successfully Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.266267 4960 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.266325 4960 factory.go:103] Registering Raw factory Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.266361 4960 manager.go:1196] Started watching for new ooms in manager Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.267569 4960 manager.go:319] Starting recovery of all containers Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.271883 4960 server.go:460] "Adding debug handlers to kubelet server" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277384 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277528 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277553 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277569 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277584 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277600 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277614 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277630 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277646 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277661 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277678 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277693 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277707 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277726 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277741 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277757 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277777 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277794 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277810 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277830 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277847 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277866 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277884 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277900 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277919 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.277978 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278055 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278086 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278128 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278153 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278172 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278191 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278214 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278236 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278254 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278272 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278290 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278312 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278331 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278351 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278370 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278390 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278410 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278434 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278452 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278471 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278488 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278510 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278557 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278579 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278599 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278627 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278653 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278676 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278696 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278717 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278735 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278754 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278772 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278790 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278810 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278829 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278848 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278866 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278887 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.278906 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279667 4960 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279705 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279729 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279748 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279769 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279787 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279806 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279824 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279843 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279859 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279883 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279901 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279918 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279935 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279954 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279971 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.279987 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280005 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280024 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280040 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280056 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280071 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280084 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280097 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280110 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280215 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280229 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280244 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280262 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280276 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280291 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280305 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280317 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280330 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280350 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280362 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280376 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280389 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280403 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280424 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280440 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280453 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280499 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280518 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280533 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280550 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280573 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280588 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280604 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280618 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280631 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280643 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280656 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280668 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280680 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280696 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280709 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280720 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280732 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280744 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280755 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280767 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280781 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280794 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280805 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280817 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280829 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280842 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280852 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280866 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280877 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280891 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280901 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280912 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280924 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280935 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280946 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280958 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280971 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280982 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.280993 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281006 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281018 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281028 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281039 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281050 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281062 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281073 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281083 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281094 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281104 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281134 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281148 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281165 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281178 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281188 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281201 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281214 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281226 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281238 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281249 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281261 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281272 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281284 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281296 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281309 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281321 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281333 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281345 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281357 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281371 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281382 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281394 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281405 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281418 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281453 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281467 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281479 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281491 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281505 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281518 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281531 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281542 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281554 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281565 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281579 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281592 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281603 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281616 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281629 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281640 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281653 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281666 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281677 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281689 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281700 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281711 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281722 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281731 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281742 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281754 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281768 4960 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281781 4960 reconstruct.go:97] "Volume reconstruction finished" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.281789 4960 reconciler.go:26] "Reconciler: start to sync state" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.301495 4960 manager.go:324] Recovery completed Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.316018 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.319305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.319373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.319385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.319486 4960 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322262 4960 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322283 4960 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322309 4960 state_mem.go:36] "Initialized new in-memory state store" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322722 4960 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322806 4960 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.322849 4960 kubelet.go:2335] "Starting kubelet main sync loop" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.322957 4960 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.325011 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.325086 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.332617 4960 policy_none.go:49] "None policy: Start" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.333517 4960 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.333643 4960 state_mem.go:35] "Initializing new in-memory state store" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.358793 4960 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411035 4960 manager.go:334] "Starting Device Plugin manager" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411139 4960 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411159 4960 server.go:79] "Starting device plugin registration server" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411727 4960 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411753 4960 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.411942 4960 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.412091 4960 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.412130 4960 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.423696 4960 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.423839 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.427310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.427347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.427360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.427539 4960 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.427561 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.429507 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.429593 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.430904 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.430931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.430940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431065 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431606 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431696 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.431875 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.432441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.432479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.432488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.432582 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.433005 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.433075 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.433844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.433890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.433904 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434192 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434298 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434564 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434653 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.434679 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.436738 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.437597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.437619 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.437627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.464021 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="400ms" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483603 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483714 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483779 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483875 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483943 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.483995 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484131 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484228 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484285 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484328 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.484360 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.512144 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.514166 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.514249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.514287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.514382 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.515055 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585558 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585630 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585666 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585785 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585834 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585880 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585947 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585959 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585922 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585970 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586006 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586050 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.585900 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586101 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586186 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586200 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586264 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586278 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586326 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586355 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586372 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586373 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586451 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586411 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586510 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.586748 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.715684 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.717461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.717508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.717527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.717560 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.718152 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.782262 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.792796 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.815219 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.838005 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5aff849218b82d411a1e90b63bc9fae05d0a7c285019cc8af75dac7075311c1e WatchSource:0}: Error finding container 5aff849218b82d411a1e90b63bc9fae05d0a7c285019cc8af75dac7075311c1e: Status 404 returned error can't find the container with id 5aff849218b82d411a1e90b63bc9fae05d0a7c285019cc8af75dac7075311c1e Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.840776 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c5518e3d8f137ebe57ec1e328941372c567a4c1d19cda630e35a88fc76b10b77 WatchSource:0}: Error finding container c5518e3d8f137ebe57ec1e328941372c567a4c1d19cda630e35a88fc76b10b77: Status 404 returned error can't find the container with id c5518e3d8f137ebe57ec1e328941372c567a4c1d19cda630e35a88fc76b10b77 Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.845469 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: I1201 19:39:25.850581 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.852784 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7fcbff0933d8b19e0f413dd6d3feb715bbfcacb829067739fcdca84da52632cb WatchSource:0}: Error finding container 7fcbff0933d8b19e0f413dd6d3feb715bbfcacb829067739fcdca84da52632cb: Status 404 returned error can't find the container with id 7fcbff0933d8b19e0f413dd6d3feb715bbfcacb829067739fcdca84da52632cb Dec 01 19:39:25 crc kubenswrapper[4960]: E1201 19:39:25.865746 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="800ms" Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.883963 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-73b3388fde89a5545f567abe53b411a7b9f3d167ac3b6b1e270c391cd843afc3 WatchSource:0}: Error finding container 73b3388fde89a5545f567abe53b411a7b9f3d167ac3b6b1e270c391cd843afc3: Status 404 returned error can't find the container with id 73b3388fde89a5545f567abe53b411a7b9f3d167ac3b6b1e270c391cd843afc3 Dec 01 19:39:25 crc kubenswrapper[4960]: W1201 19:39:25.889717 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-652be7e2a4ddc7635e126a0a55c816712591ddafae92da1e306256a190c3afd9 WatchSource:0}: Error finding container 652be7e2a4ddc7635e126a0a55c816712591ddafae92da1e306256a190c3afd9: Status 404 returned error can't find the container with id 652be7e2a4ddc7635e126a0a55c816712591ddafae92da1e306256a190c3afd9 Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.119099 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.121228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.121288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.121307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.121347 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.122008 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.253365 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.327762 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.327918 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"73b3388fde89a5545f567abe53b411a7b9f3d167ac3b6b1e270c391cd843afc3"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.331000 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.331057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7fcbff0933d8b19e0f413dd6d3feb715bbfcacb829067739fcdca84da52632cb"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.331197 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.332618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.332642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.332652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.333841 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.333874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5aff849218b82d411a1e90b63bc9fae05d0a7c285019cc8af75dac7075311c1e"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.333955 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.334760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.334780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.334790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.336962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.336992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c5518e3d8f137ebe57ec1e328941372c567a4c1d19cda630e35a88fc76b10b77"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.337076 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.337889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.337936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.337949 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339029 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339059 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"652be7e2a4ddc7635e126a0a55c816712591ddafae92da1e306256a190c3afd9"} Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339186 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339763 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.339773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: W1201 19:39:26.508150 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.508280 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.666774 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="1.6s" Dec 01 19:39:26 crc kubenswrapper[4960]: W1201 19:39:26.714618 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.714754 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:26 crc kubenswrapper[4960]: W1201 19:39:26.737370 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.737549 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:26 crc kubenswrapper[4960]: W1201 19:39:26.814321 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.814419 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.177:6443: connect: connection refused" logger="UnhandledError" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.922980 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.924452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.924505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.924521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:26 crc kubenswrapper[4960]: I1201 19:39:26.924554 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:26 crc kubenswrapper[4960]: E1201 19:39:26.925239 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.177:6443: connect: connection refused" node="crc" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.252567 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.343141 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb" exitCode=0 Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.343192 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.343469 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.344837 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.344880 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.344897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.345449 4960 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2" exitCode=0 Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.345534 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.345607 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.350530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.350680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.350767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.355013 4960 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87" exitCode=0 Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.355197 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.355346 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.357582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.357628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.357638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.359255 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.359317 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.359331 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.359290 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.360218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.360249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.360260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.361532 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b" exitCode=0 Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.361579 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b"} Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.361683 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.362434 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.362464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.362474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.368639 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.369334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.369365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:27 crc kubenswrapper[4960]: I1201 19:39:27.369375 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.252316 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.177:6443: connect: connection refused Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.368471 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5" exitCode=0 Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.368555 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.368762 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.370617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.370647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.370658 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.374324 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c36c9bc6acdea251df7b1885d7643ce119ef3535f7ba3d17986090471d91f1ad"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.374575 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.375746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.375780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.375792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.382998 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.383045 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.383062 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.383200 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.384560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.384586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.384599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.387732 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388094 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388149 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388166 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388179 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2"} Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388513 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.388529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.526250 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.531292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.531364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.531383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:28 crc kubenswrapper[4960]: I1201 19:39:28.531431 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.396628 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a"} Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.396767 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.398584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.398647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.398676 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.402188 4960 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b" exitCode=0 Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.402256 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b"} Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.402491 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.403861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.403929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.403954 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.404014 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.592094 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.592358 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.594193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.594265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.594281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.746768 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.754518 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:29 crc kubenswrapper[4960]: I1201 19:39:29.824856 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.308419 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.408752 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610"} Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.408822 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.408841 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a"} Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.408877 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5"} Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.408906 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410751 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410928 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:30 crc kubenswrapper[4960]: I1201 19:39:30.410956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.417958 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba"} Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.418037 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a"} Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.418071 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.418204 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.418261 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419615 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419886 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419932 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.419925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.420181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.420206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.841004 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:31 crc kubenswrapper[4960]: I1201 19:39:31.940552 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.120754 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.121022 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.128109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.128246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.128270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.420475 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.420480 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421811 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421880 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.421953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.592973 4960 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 19:39:32 crc kubenswrapper[4960]: I1201 19:39:32.593080 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 19:39:33 crc kubenswrapper[4960]: I1201 19:39:33.423190 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:33 crc kubenswrapper[4960]: I1201 19:39:33.424528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:33 crc kubenswrapper[4960]: I1201 19:39:33.424602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:33 crc kubenswrapper[4960]: I1201 19:39:33.424624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:34 crc kubenswrapper[4960]: I1201 19:39:34.322232 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:34 crc kubenswrapper[4960]: I1201 19:39:34.322476 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:34 crc kubenswrapper[4960]: I1201 19:39:34.323891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:34 crc kubenswrapper[4960]: I1201 19:39:34.323933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:34 crc kubenswrapper[4960]: I1201 19:39:34.323946 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:35 crc kubenswrapper[4960]: E1201 19:39:35.427740 4960 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 01 19:39:38 crc kubenswrapper[4960]: E1201 19:39:38.268638 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 01 19:39:38 crc kubenswrapper[4960]: E1201 19:39:38.532730 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 01 19:39:38 crc kubenswrapper[4960]: W1201 19:39:38.544059 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 19:39:38 crc kubenswrapper[4960]: I1201 19:39:38.544218 4960 trace.go:236] Trace[1338775666]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 19:39:28.542) (total time: 10001ms): Dec 01 19:39:38 crc kubenswrapper[4960]: Trace[1338775666]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:39:38.544) Dec 01 19:39:38 crc kubenswrapper[4960]: Trace[1338775666]: [10.001404311s] [10.001404311s] END Dec 01 19:39:38 crc kubenswrapper[4960]: E1201 19:39:38.544265 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 19:39:39 crc kubenswrapper[4960]: W1201 19:39:39.209266 4960 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.209408 4960 trace.go:236] Trace[964299653]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 19:39:29.207) (total time: 10001ms): Dec 01 19:39:39 crc kubenswrapper[4960]: Trace[964299653]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (19:39:39.209) Dec 01 19:39:39 crc kubenswrapper[4960]: Trace[964299653]: [10.001607158s] [10.001607158s] END Dec 01 19:39:39 crc kubenswrapper[4960]: E1201 19:39:39.209448 4960 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.254016 4960 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.488352 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.488444 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.505973 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.506074 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.833950 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]log ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]etcd ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-filter ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-apiextensions-informers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-apiextensions-controllers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/crd-informer-synced ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-system-namespaces-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 01 19:39:39 crc kubenswrapper[4960]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/bootstrap-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/start-kube-aggregator-informers ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-registration-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-discovery-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]autoregister-completion ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-openapi-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 01 19:39:39 crc kubenswrapper[4960]: livez check failed Dec 01 19:39:39 crc kubenswrapper[4960]: I1201 19:39:39.834034 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:39:40 crc kubenswrapper[4960]: I1201 19:39:40.315850 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:40 crc kubenswrapper[4960]: I1201 19:39:40.316023 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:40 crc kubenswrapper[4960]: I1201 19:39:40.317229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:40 crc kubenswrapper[4960]: I1201 19:39:40.317265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:40 crc kubenswrapper[4960]: I1201 19:39:40.317278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.106336 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.106669 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.108309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.108361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.108379 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.140888 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.445978 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.447420 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.447506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.447533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.468179 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.733031 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.735109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.735172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.735181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:41 crc kubenswrapper[4960]: I1201 19:39:41.735210 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:41 crc kubenswrapper[4960]: E1201 19:39:41.741689 4960 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.448696 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.449738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.449791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.449806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.592710 4960 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.592825 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 01 19:39:42 crc kubenswrapper[4960]: I1201 19:39:42.663411 4960 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.229662 4960 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.254337 4960 apiserver.go:52] "Watching apiserver" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.258963 4960 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.259751 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.260589 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.260879 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:43 crc kubenswrapper[4960]: E1201 19:39:43.262074 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.264965 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.265736 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.265913 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:43 crc kubenswrapper[4960]: E1201 19:39:43.265920 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.266214 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:43 crc kubenswrapper[4960]: E1201 19:39:43.267490 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.267755 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.268341 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.269329 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.269891 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.271609 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.271742 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.271910 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.271959 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.272991 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.317076 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.346006 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.358741 4960 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.365892 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.382235 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.394815 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.410074 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:43 crc kubenswrapper[4960]: I1201 19:39:43.423918 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.479986 4960 trace.go:236] Trace[906673755]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 19:39:29.521) (total time: 14958ms): Dec 01 19:39:44 crc kubenswrapper[4960]: Trace[906673755]: ---"Objects listed" error: 14958ms (19:39:44.479) Dec 01 19:39:44 crc kubenswrapper[4960]: Trace[906673755]: [14.958320804s] [14.958320804s] END Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.480069 4960 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.480077 4960 trace.go:236] Trace[541737027]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (01-Dec-2025 19:39:29.928) (total time: 14551ms): Dec 01 19:39:44 crc kubenswrapper[4960]: Trace[541737027]: ---"Objects listed" error: 14551ms (19:39:44.479) Dec 01 19:39:44 crc kubenswrapper[4960]: Trace[541737027]: [14.55165168s] [14.55165168s] END Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.480160 4960 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.482142 4960 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.519935 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38024->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.519981 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38028->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.520036 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38024->192.168.126.11:17697: read: connection reset by peer" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.520053 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38028->192.168.126.11:17697: read: connection reset by peer" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583582 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583767 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583820 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583849 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583870 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583891 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583921 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583956 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583977 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.583995 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584016 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584057 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584086 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584134 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584188 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584211 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584233 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584260 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584281 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584325 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584363 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584391 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584402 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584437 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584544 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584581 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584613 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584639 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584663 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584693 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584714 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584731 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584749 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584771 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584791 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584808 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584827 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584845 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584862 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584879 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584896 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584920 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.584980 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585006 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585061 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585078 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585099 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585127 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585144 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585176 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585203 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585227 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585223 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585251 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585251 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585302 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585323 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585351 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585351 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585427 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585477 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585516 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585552 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585575 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585586 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585606 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585599 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585652 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585681 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585709 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585760 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585792 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585801 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585825 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585838 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585822 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585891 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585866 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585981 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.585981 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586029 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586055 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586070 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586077 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586135 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586132 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586158 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586192 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586222 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586256 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586263 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586284 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586308 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586360 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586383 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586402 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586419 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586438 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586456 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586479 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586496 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586513 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586531 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586548 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586564 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586579 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586599 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586617 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586633 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586649 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586668 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586685 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586721 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586738 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586755 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586772 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586793 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586811 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586829 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586845 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586863 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586881 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586896 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586912 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586930 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586967 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586985 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587049 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587066 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587085 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587103 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587137 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587155 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587178 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587200 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587217 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587237 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587253 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587270 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587285 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587320 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587336 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587353 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587371 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587389 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587406 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587424 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587441 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587461 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587478 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587497 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587517 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587536 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587556 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587574 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587595 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587615 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587639 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587655 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587675 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587707 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587730 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587750 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587768 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587785 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587805 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587824 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587845 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587863 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587880 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587897 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587916 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587934 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587955 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587972 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587989 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588007 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588051 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588068 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588085 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588104 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588840 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588869 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588907 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588924 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588943 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588960 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588977 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588993 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589012 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589028 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589053 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589071 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589210 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589236 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589256 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589274 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589321 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589367 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589388 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589411 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589432 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589450 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589470 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589514 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589534 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589556 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589617 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589720 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589736 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589749 4960 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589764 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589778 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589792 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589810 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589821 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589832 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589842 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589851 4960 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589861 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589874 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589885 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589895 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589905 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589914 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586384 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586440 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586479 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586594 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586763 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586809 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586898 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.586939 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587151 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587150 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587544 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587797 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587811 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.587837 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.588990 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589558 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589666 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589776 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.589946 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590078 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590108 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590340 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590408 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590419 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590430 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590572 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590762 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590872 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.590965 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591052 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591264 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591293 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591301 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591314 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591629 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591682 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591898 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.591971 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.592225 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.592436 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595054 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595076 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.599225 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595416 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595475 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595637 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.595936 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.596164 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.596351 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.596534 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.596828 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.597400 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.597580 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.597772 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.598143 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.599414 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.599423 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600045 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600134 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600348 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600369 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600621 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600737 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600810 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600889 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.600983 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.601039 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.601083 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.601176 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.601847 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.601859 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602258 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602287 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602361 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602671 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602655 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602788 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.602978 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.603008 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.603881 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.603884 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604051 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604132 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604295 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604390 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604423 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604494 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.604613 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:39:45.104509409 +0000 UTC m=+20.392001078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604633 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604685 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604763 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604848 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.604928 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.605245 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.605279 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.605727 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.605828 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606166 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606443 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.606588 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606611 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.606649 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:45.106627895 +0000 UTC m=+20.394119564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606664 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606655 4960 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.606742 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.606811 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:45.10680147 +0000 UTC m=+20.394293339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606863 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606905 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606920 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.606940 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.607264 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.607590 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608164 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608226 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608540 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608548 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608620 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.608736 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.609372 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.609758 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.609908 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.609965 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.609740 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.610202 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.612741 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613268 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613470 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613566 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613589 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613786 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.613899 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.617281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.617517 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.617562 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.618154 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.618527 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.618846 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.619243 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.621105 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.621273 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.621293 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.621307 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.621362 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:45.121345776 +0000 UTC m=+20.408837445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.621456 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.623199 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.623215 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.623340 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.623740 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.623928 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.624727 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.624750 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.624763 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:44 crc kubenswrapper[4960]: E1201 19:39:44.624801 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:45.124791242 +0000 UTC m=+20.412282911 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.625521 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.625748 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.629525 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.630755 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.630845 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.631991 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.631949 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.636489 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.636595 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.637079 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.637195 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.637503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.637612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.637637 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638096 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638176 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638892 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638979 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.638953 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639357 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639425 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639401 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.640231 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.640305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639669 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639750 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.639815 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.640159 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.640380 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.640872 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.641461 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.641299 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.662627 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.680366 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.686260 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690489 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690565 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690648 4960 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690663 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690676 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690689 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690703 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690717 4960 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690731 4960 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690745 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690759 4960 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690772 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690784 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690797 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690808 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690822 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690834 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690848 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690860 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690872 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690884 4960 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690896 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690908 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690920 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690932 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690944 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690956 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690979 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.690991 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691003 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691015 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691026 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691038 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691050 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691062 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691073 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691085 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691098 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691125 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691137 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691149 4960 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691162 4960 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691174 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691186 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691199 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691212 4960 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691225 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691238 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691249 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691262 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691274 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691286 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691298 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691313 4960 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691326 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691339 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691351 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691363 4960 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691375 4960 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691387 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691398 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691412 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691425 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691436 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691448 4960 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691460 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691483 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691494 4960 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691505 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691517 4960 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691529 4960 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691541 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691552 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691564 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691576 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691587 4960 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691601 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691613 4960 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691624 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691636 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691647 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691660 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691672 4960 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691684 4960 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691695 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691706 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691719 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691731 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691742 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691754 4960 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691766 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691777 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691789 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691802 4960 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691814 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691824 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691835 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691850 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691862 4960 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691874 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691887 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691898 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691910 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691922 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691933 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691945 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691956 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691968 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691979 4960 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.691991 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692003 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692016 4960 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692029 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692041 4960 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692052 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692065 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692077 4960 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692088 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692131 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692143 4960 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692155 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692168 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692181 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692193 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692204 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692216 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692227 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692209 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692239 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692314 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692329 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692344 4960 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692356 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692369 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692383 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692396 4960 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692406 4960 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692417 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692429 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692440 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692356 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692475 4960 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692491 4960 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692503 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692514 4960 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692528 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692540 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692553 4960 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692565 4960 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692577 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692589 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692601 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692614 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692627 4960 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692642 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692656 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692671 4960 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692690 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692702 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692714 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692729 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692742 4960 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692754 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692767 4960 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692781 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692792 4960 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692805 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692818 4960 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692830 4960 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692842 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692855 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692866 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692878 4960 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692889 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692900 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692911 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692922 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692935 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692946 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692957 4960 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.692968 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.791760 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.806554 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 01 19:39:44 crc kubenswrapper[4960]: W1201 19:39:44.807367 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ccd5ecd12c3235c8a8490c4fb16b81ad0733254906a70b248f9d699cb69a1466 WatchSource:0}: Error finding container ccd5ecd12c3235c8a8490c4fb16b81ad0733254906a70b248f9d699cb69a1466: Status 404 returned error can't find the container with id ccd5ecd12c3235c8a8490c4fb16b81ad0733254906a70b248f9d699cb69a1466 Dec 01 19:39:44 crc kubenswrapper[4960]: W1201 19:39:44.819673 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-3f83a9b7ea151d24e8afe6b92f46722b4a1ee03a11aa33820a230f60bbe6e5ca WatchSource:0}: Error finding container 3f83a9b7ea151d24e8afe6b92f46722b4a1ee03a11aa33820a230f60bbe6e5ca: Status 404 returned error can't find the container with id 3f83a9b7ea151d24e8afe6b92f46722b4a1ee03a11aa33820a230f60bbe6e5ca Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.822031 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.829635 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.830393 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.830483 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.835664 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:44 crc kubenswrapper[4960]: W1201 19:39:44.840806 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-95ffa6e0aaf9fe62dcab765fe0836ed1be2a6e0a417821c17e90de4ba605d7b1 WatchSource:0}: Error finding container 95ffa6e0aaf9fe62dcab765fe0836ed1be2a6e0a417821c17e90de4ba605d7b1: Status 404 returned error can't find the container with id 95ffa6e0aaf9fe62dcab765fe0836ed1be2a6e0a417821c17e90de4ba605d7b1 Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.841776 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.843833 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.855298 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.867428 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.880670 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.893167 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.904924 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.936588 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.969830 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.983212 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:44 crc kubenswrapper[4960]: I1201 19:39:44.994935 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.007050 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.018233 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.031580 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.196527 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.196596 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.196623 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.196641 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.196663 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196737 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:39:46.196717671 +0000 UTC m=+21.484209340 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196773 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196809 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196826 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196856 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:46.196836925 +0000 UTC m=+21.484328594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196826 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196898 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196910 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196923 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:46.196900177 +0000 UTC m=+21.484391846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196833 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196944 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:46.196936088 +0000 UTC m=+21.484427757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196952 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.196977 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:46.196970229 +0000 UTC m=+21.484461898 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.324046 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.324134 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.324086 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.324272 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.324295 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.324429 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.327844 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.328685 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.329589 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.330376 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.330972 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.331613 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.332302 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.332902 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.333705 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.334142 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.335087 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.335606 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.336291 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.336819 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.337333 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.337825 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.338408 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.338990 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.339395 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.339943 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.340554 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.341029 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.341646 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.342091 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.342741 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.342988 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.343206 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.343797 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.344422 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.348133 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.348802 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.349838 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.350428 4960 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.350551 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.352730 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.353389 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.353841 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.355326 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.356060 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.357292 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.357861 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.358559 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.359399 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.359945 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.361432 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.362083 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.363087 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.363586 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.364475 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.364952 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.366063 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.366536 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.367389 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.367829 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.368325 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.369243 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.369688 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.372757 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.383326 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.398924 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.406687 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.458197 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.459657 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a" exitCode=255 Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.459712 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.460775 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"95ffa6e0aaf9fe62dcab765fe0836ed1be2a6e0a417821c17e90de4ba605d7b1"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.462534 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.462588 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.462603 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3f83a9b7ea151d24e8afe6b92f46722b4a1ee03a11aa33820a230f60bbe6e5ca"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.463922 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815"} Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.463950 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ccd5ecd12c3235c8a8490c4fb16b81ad0733254906a70b248f9d699cb69a1466"} Dec 01 19:39:45 crc kubenswrapper[4960]: E1201 19:39:45.469800 4960 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.470063 4960 scope.go:117] "RemoveContainer" containerID="8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.486334 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.500827 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.513638 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.541566 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.554759 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.567778 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.583592 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.595815 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.610494 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.625999 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.638748 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.653197 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.671240 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:45 crc kubenswrapper[4960]: I1201 19:39:45.689836 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.208594 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.208701 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.208748 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.208776 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.208809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.208966 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209008 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209058 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209070 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209137 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209157 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209166 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209082 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209035 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:48.209015418 +0000 UTC m=+23.496507107 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209282 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:48.209254515 +0000 UTC m=+23.496746394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209306 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:48.209294586 +0000 UTC m=+23.496786515 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209371 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:39:48.209339378 +0000 UTC m=+23.496831087 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:39:46 crc kubenswrapper[4960]: E1201 19:39:46.209405 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:48.209391089 +0000 UTC m=+23.496882798 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.469320 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.471437 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1"} Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.471854 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.487741 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.506627 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.520012 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.532429 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.548553 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.567828 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:46 crc kubenswrapper[4960]: I1201 19:39:46.610294 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:47 crc kubenswrapper[4960]: I1201 19:39:47.323901 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:47 crc kubenswrapper[4960]: I1201 19:39:47.323948 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:47 crc kubenswrapper[4960]: I1201 19:39:47.323909 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:47 crc kubenswrapper[4960]: E1201 19:39:47.324054 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:47 crc kubenswrapper[4960]: E1201 19:39:47.324189 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:47 crc kubenswrapper[4960]: E1201 19:39:47.324300 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.142310 4960 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.144918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.145002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.145014 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.145133 4960 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.154951 4960 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.155456 4960 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.156976 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.157022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.157032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.157051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.157061 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.197965 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.202065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.202126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.202135 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.202149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.202159 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.224864 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.224991 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.225026 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.225053 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225083 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:39:52.225050819 +0000 UTC m=+27.512542488 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225155 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.225154 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225223 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225245 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225306 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225318 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225252 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225380 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225225 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:52.225205254 +0000 UTC m=+27.512697103 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225427 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:52.22540859 +0000 UTC m=+27.512900259 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225441 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:52.225435141 +0000 UTC m=+27.512926810 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225262 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.225504 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:39:52.225492943 +0000 UTC m=+27.512984612 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.235623 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.245681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.245712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.245719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.245732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.245741 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.270164 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.279596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.279644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.279654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.279669 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.279679 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.304788 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.310624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.310672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.310681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.310695 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.310704 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.333964 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: E1201 19:39:48.334083 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.335719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.335755 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.335764 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.335783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.335796 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.438698 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.438756 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.438770 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.438789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.438807 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.479208 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.498389 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.516014 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.533461 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.541175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.541233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.541244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.541262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.541280 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.548599 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.566199 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.583992 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.601822 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.626362 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lgrxx"] Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.626755 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.629176 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.629346 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.629435 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.643522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.643573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.643585 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.643605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.643620 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.645317 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.659311 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.683900 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.702730 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.720772 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.729128 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-hosts-file\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.729181 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-628zd\" (UniqueName: \"kubernetes.io/projected/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-kube-api-access-628zd\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.740104 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.746207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.746265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.746278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.746306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.746321 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.760923 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.776524 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.830213 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-628zd\" (UniqueName: \"kubernetes.io/projected/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-kube-api-access-628zd\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.830308 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-hosts-file\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.830428 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-hosts-file\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.849403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.849445 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.849456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.849473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.849484 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.867016 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-628zd\" (UniqueName: \"kubernetes.io/projected/c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471-kube-api-access-628zd\") pod \"node-resolver-lgrxx\" (UID: \"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\") " pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.942179 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lgrxx" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.952755 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.952809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.952819 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.952837 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:48 crc kubenswrapper[4960]: I1201 19:39:48.952880 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:48Z","lastTransitionTime":"2025-12-01T19:39:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.002405 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ct7db"] Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.003241 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.005565 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-gns76"] Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.006034 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.006704 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-snl4j"] Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.008970 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.014810 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.015056 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.015140 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.015157 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.015379 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.019071 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.019231 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.019392 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.019616 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.020425 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.021134 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.021268 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.032123 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.032160 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crklw\" (UniqueName: \"kubernetes.io/projected/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-kube-api-access-crklw\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.032205 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-rootfs\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.032333 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-proxy-tls\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.038375 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.056571 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.062422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.062480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.062490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.062508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.062518 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.071323 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.087486 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.101637 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.119046 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.133937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-k8s-cni-cncf-io\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.133995 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134023 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-system-cni-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134045 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134070 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmtp8\" (UniqueName: \"kubernetes.io/projected/804735e5-ed5b-4c59-8801-02f492e86d4e-kube-api-access-gmtp8\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-bin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134146 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-conf-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134170 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk8tc\" (UniqueName: \"kubernetes.io/projected/cfba83a1-281b-4f8f-be15-2760b2b9dfce-kube-api-access-sk8tc\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134203 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-cnibin\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-kubelet\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134249 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cnibin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134278 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cni-binary-copy\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134305 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-rootfs\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134335 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-etc-kubernetes\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134358 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134387 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-multus-certs\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134411 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-system-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134436 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134463 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crklw\" (UniqueName: \"kubernetes.io/projected/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-kube-api-access-crklw\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134495 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-netns\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134521 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-os-release\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134567 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-multus\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134590 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-daemon-config\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134613 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-socket-dir-parent\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134637 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-proxy-tls\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134658 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-hostroot\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134690 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.134718 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-os-release\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.135531 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.135688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-rootfs\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.136816 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.141388 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-proxy-tls\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.158990 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.161484 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crklw\" (UniqueName: \"kubernetes.io/projected/b6dbabf7-fd52-4f8d-9bca-093018d1c0b9-kube-api-access-crklw\") pod \"machine-config-daemon-ct7db\" (UID: \"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\") " pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.165488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.165540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.165553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.165569 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.165579 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.171624 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.186358 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.200051 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.213741 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.227813 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236147 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-netns\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236199 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-os-release\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236240 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-multus\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236260 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-daemon-config\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236282 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-socket-dir-parent\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-hostroot\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-netns\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236392 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-multus\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236333 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236512 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-os-release\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236527 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-os-release\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236564 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-socket-dir-parent\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236591 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-hostroot\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-k8s-cni-cncf-io\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236614 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-k8s-cni-cncf-io\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236639 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-system-cni-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236658 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236677 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-os-release\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmtp8\" (UniqueName: \"kubernetes.io/projected/804735e5-ed5b-4c59-8801-02f492e86d4e-kube-api-access-gmtp8\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236712 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-cnibin\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236747 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-bin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236765 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-conf-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk8tc\" (UniqueName: \"kubernetes.io/projected/cfba83a1-281b-4f8f-be15-2760b2b9dfce-kube-api-access-sk8tc\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236796 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cni-binary-copy\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-kubelet\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cnibin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236885 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-etc-kubernetes\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236904 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236928 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-multus-certs\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.236984 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-system-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237000 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237065 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-system-cni-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237070 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-daemon-config\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-kubelet\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237177 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-cnibin\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237213 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-var-lib-cni-bin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237244 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-conf-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-multus-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237393 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cnibin\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237479 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-host-run-multus-certs\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237535 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-system-cni-dir\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237596 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfba83a1-281b-4f8f-be15-2760b2b9dfce-etc-kubernetes\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.237871 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-binary-copy\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.238025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/804735e5-ed5b-4c59-8801-02f492e86d4e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.238153 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/804735e5-ed5b-4c59-8801-02f492e86d4e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.238336 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cfba83a1-281b-4f8f-be15-2760b2b9dfce-cni-binary-copy\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.246290 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.259359 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmtp8\" (UniqueName: \"kubernetes.io/projected/804735e5-ed5b-4c59-8801-02f492e86d4e-kube-api-access-gmtp8\") pod \"multus-additional-cni-plugins-snl4j\" (UID: \"804735e5-ed5b-4c59-8801-02f492e86d4e\") " pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.260050 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk8tc\" (UniqueName: \"kubernetes.io/projected/cfba83a1-281b-4f8f-be15-2760b2b9dfce-kube-api-access-sk8tc\") pod \"multus-gns76\" (UID: \"cfba83a1-281b-4f8f-be15-2760b2b9dfce\") " pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.260577 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.268097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.268156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.268167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.268184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.268197 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.275811 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.290581 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.302880 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.316496 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.323731 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.323826 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.323888 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:49 crc kubenswrapper[4960]: E1201 19:39:49.323884 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:49 crc kubenswrapper[4960]: E1201 19:39:49.324012 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:49 crc kubenswrapper[4960]: E1201 19:39:49.324094 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.325816 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.331048 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.337679 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gns76" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.349947 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-snl4j" Dec 01 19:39:49 crc kubenswrapper[4960]: W1201 19:39:49.351806 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfba83a1_281b_4f8f_be15_2760b2b9dfce.slice/crio-80b5643365546171601ef3713d567c08227183d9c4c0c57e2da337bd78fa89da WatchSource:0}: Error finding container 80b5643365546171601ef3713d567c08227183d9c4c0c57e2da337bd78fa89da: Status 404 returned error can't find the container with id 80b5643365546171601ef3713d567c08227183d9c4c0c57e2da337bd78fa89da Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.372054 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.372126 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.372142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.372165 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.372179 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.387279 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fdntq"] Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.388377 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.393910 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.394032 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.394128 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.395860 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.396598 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.396785 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.398485 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.435668 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440265 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440299 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440336 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440352 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440367 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440389 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440405 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440423 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440443 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440493 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbhw\" (UniqueName: \"kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440649 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440699 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440723 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440743 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440759 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440789 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440824 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440841 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.440860 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.452607 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.473194 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.475447 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.475583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.475598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.475620 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.475652 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.489905 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.489974 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"5e348a08114e36e6ce6c71355732daa35bacd8cecc55df2a5cc8d0a4654fb454"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.497686 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.502724 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerStarted","Data":"717ba79a65150217f7cc3c3c06b853a22f014ad38f569499fb7c7f901d02b183"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.506447 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerStarted","Data":"80b5643365546171601ef3713d567c08227183d9c4c0c57e2da337bd78fa89da"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.518493 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.518994 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lgrxx" event={"ID":"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471","Type":"ContainerStarted","Data":"10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.519100 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lgrxx" event={"ID":"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471","Type":"ContainerStarted","Data":"8d8e8be426787096db307e3424552efa94e0251cb9cb5da66d7362882be1b86a"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.544424 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.544707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.544852 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545056 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545089 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545129 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545132 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545227 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545250 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545198 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545231 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545296 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545311 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545362 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545395 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545398 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545437 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545457 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbhw\" (UniqueName: \"kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545479 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545497 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545515 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545543 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545590 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545617 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545661 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545686 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545780 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545810 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.545972 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546237 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546337 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546439 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546687 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.546810 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.547443 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.558815 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.564473 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.571572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbhw\" (UniqueName: \"kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw\") pod \"ovnkube-node-fdntq\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.577864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.578038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.578132 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.578207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.578267 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.588845 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.596072 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.599598 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.602244 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.604542 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.614293 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.626291 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.647496 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.667459 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.680989 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.681047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.681062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.681087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.681100 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.683652 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.696497 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.711511 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.727935 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.728987 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.741300 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: W1201 19:39:49.741857 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c9203b1_b24d_40f4_84e1_180293ad742b.slice/crio-2cc943a224c9bac2a59edaefe4f1128741dd2f219f9e6194b54aec9ccb8e2c82 WatchSource:0}: Error finding container 2cc943a224c9bac2a59edaefe4f1128741dd2f219f9e6194b54aec9ccb8e2c82: Status 404 returned error can't find the container with id 2cc943a224c9bac2a59edaefe4f1128741dd2f219f9e6194b54aec9ccb8e2c82 Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.756621 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.777310 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.784310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.784369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.784383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.784404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.784417 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.790163 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.806587 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.824624 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.838204 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.854212 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.887478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.887540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.887559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.887583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.887601 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.990894 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.991321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.991336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.991355 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:49 crc kubenswrapper[4960]: I1201 19:39:49.991370 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:49Z","lastTransitionTime":"2025-12-01T19:39:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.093996 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.094028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.094038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.094053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.094063 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.198226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.198260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.198270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.198284 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.198296 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.301501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.301560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.301579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.301602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.301619 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.404938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.405429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.405439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.405454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.405463 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.508388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.508450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.508469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.508494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.508512 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.524915 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" exitCode=0 Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.525040 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.525167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"2cc943a224c9bac2a59edaefe4f1128741dd2f219f9e6194b54aec9ccb8e2c82"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.527399 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84" exitCode=0 Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.527470 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.532102 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.537819 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerStarted","Data":"e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.547937 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.565559 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.588340 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.608648 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.612807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.612848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.612867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.612884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.612894 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.623588 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.638985 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.658928 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.675588 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.691926 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.708345 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.719346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.719398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.719414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.719435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.719455 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.726241 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.745323 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.769055 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.787603 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.801174 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.814805 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.823189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.823243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.823258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.823277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.823288 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.830484 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.843873 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.854830 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.867238 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.883679 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.898017 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.911921 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.926279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.926339 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.926353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.926374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.926389 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:50Z","lastTransitionTime":"2025-12-01T19:39:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.934958 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.952294 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:50 crc kubenswrapper[4960]: I1201 19:39:50.968597 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:50Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.028727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.028762 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.028773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.028815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.028827 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.132539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.132605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.132627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.132653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.132672 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.235072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.235145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.235161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.235181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.235193 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.323677 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.323789 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:51 crc kubenswrapper[4960]: E1201 19:39:51.323823 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.323677 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:51 crc kubenswrapper[4960]: E1201 19:39:51.323998 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:51 crc kubenswrapper[4960]: E1201 19:39:51.324048 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.338237 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.338278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.338290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.338311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.338325 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.441953 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.442021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.442060 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.442082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.442094 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.545992 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.546044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.546063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.546096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.546137 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552397 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552523 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552554 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552579 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552605 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.552631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.555240 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb" exitCode=0 Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.555358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.573731 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.590462 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.618994 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.632286 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.657251 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.674492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.674535 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.674548 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.674567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.674580 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.688500 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.692740 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-hcrp4"] Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.693759 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.697637 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.697983 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.698101 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.698173 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.753739 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.767156 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.774268 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09fbe457-1574-4501-837a-fb6b55fe46b4-serviceca\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.774306 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glscm\" (UniqueName: \"kubernetes.io/projected/09fbe457-1574-4501-837a-fb6b55fe46b4-kube-api-access-glscm\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.774328 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09fbe457-1574-4501-837a-fb6b55fe46b4-host\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.776767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.776798 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.776808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.776822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.776849 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.780259 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.792722 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.805591 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.819213 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.831758 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.845482 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.862778 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.875171 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.875634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glscm\" (UniqueName: \"kubernetes.io/projected/09fbe457-1574-4501-837a-fb6b55fe46b4-kube-api-access-glscm\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.875736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09fbe457-1574-4501-837a-fb6b55fe46b4-serviceca\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.875817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09fbe457-1574-4501-837a-fb6b55fe46b4-host\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.875959 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09fbe457-1574-4501-837a-fb6b55fe46b4-host\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.877080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/09fbe457-1574-4501-837a-fb6b55fe46b4-serviceca\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.879829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.879860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.879872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.879887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.879897 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.888859 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.894988 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glscm\" (UniqueName: \"kubernetes.io/projected/09fbe457-1574-4501-837a-fb6b55fe46b4-kube-api-access-glscm\") pod \"node-ca-hcrp4\" (UID: \"09fbe457-1574-4501-837a-fb6b55fe46b4\") " pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.903598 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.923913 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.940364 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.953743 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.967059 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.978257 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.982010 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.982065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.982085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.982131 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.982147 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:51Z","lastTransitionTime":"2025-12-01T19:39:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:51 crc kubenswrapper[4960]: I1201 19:39:51.998943 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:51Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.011194 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.011527 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-hcrp4" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.024435 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.037663 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.086147 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.086203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.086219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.086242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.086256 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.189203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.189274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.189294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.189320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.189337 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.280568 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.280764 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:40:00.280721865 +0000 UTC m=+35.568213574 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.281194 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.281234 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.281357 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.281408 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281540 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281625 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:00.281577262 +0000 UTC m=+35.569068941 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281821 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281850 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281863 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281878 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281913 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.282010 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.281871 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.282012 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:00.281983584 +0000 UTC m=+35.569475463 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.282155 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:00.282142689 +0000 UTC m=+35.569634548 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:52 crc kubenswrapper[4960]: E1201 19:39:52.282234 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:00.282219972 +0000 UTC m=+35.569711661 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.292573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.292613 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.292624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.292641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.292654 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.395864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.395906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.395915 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.395933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.395944 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.498737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.498836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.498852 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.498871 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.498888 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.562298 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-hcrp4" event={"ID":"09fbe457-1574-4501-837a-fb6b55fe46b4","Type":"ContainerStarted","Data":"c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.562362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-hcrp4" event={"ID":"09fbe457-1574-4501-837a-fb6b55fe46b4","Type":"ContainerStarted","Data":"2825786dc991f82a00d85edc0bd3be995c54c18622138591bf4007763716802d"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.565907 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861" exitCode=0 Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.565997 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.580179 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.593440 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.601689 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.601746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.601760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.601779 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.601792 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.616570 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.635044 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.655580 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.677690 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.696451 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.705714 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.705813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.705863 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.705891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.705911 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.722238 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.737170 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.749589 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.760980 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.776734 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.790393 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.804021 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.810659 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.810708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.810723 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.810750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.810765 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.821848 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.837736 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.850530 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.864675 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.877293 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.890058 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.901020 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.913154 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.913208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.913221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.913251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.913264 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:52Z","lastTransitionTime":"2025-12-01T19:39:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.921774 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.935831 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.947842 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.961881 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.982642 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:52 crc kubenswrapper[4960]: I1201 19:39:52.998334 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:52Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.009369 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.015013 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.015044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.015055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.015074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.015087 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.117862 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.117911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.117921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.117937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.117946 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.220621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.220704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.220726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.220752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.220771 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.323252 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.323309 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:53 crc kubenswrapper[4960]: E1201 19:39:53.323434 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.323592 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:53 crc kubenswrapper[4960]: E1201 19:39:53.323815 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:53 crc kubenswrapper[4960]: E1201 19:39:53.323948 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.324781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.324849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.324868 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.324894 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.324912 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.427587 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.427675 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.427694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.427717 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.427736 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.531793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.531877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.531902 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.531933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.531957 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.574865 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d" exitCode=0 Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.574945 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.591842 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.606271 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.627532 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.634474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.634510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.634526 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.634547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.634565 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.647645 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.661159 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.673873 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.687449 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.701108 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.724139 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.736883 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.737085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.737112 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.737138 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.737155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.737168 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.755750 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.770248 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.781869 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.795754 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:53Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.839533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.839569 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.839581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.839599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.839611 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.942636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.942713 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.942728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.942752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:53 crc kubenswrapper[4960]: I1201 19:39:53.942783 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:53Z","lastTransitionTime":"2025-12-01T19:39:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.045804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.045864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.045879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.045905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.045923 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.149350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.149411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.149422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.149441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.149459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.254061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.254143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.254161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.254184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.254200 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.358182 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.358934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.358987 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.359017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.359040 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.463566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.463642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.463660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.463690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.463709 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.566860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.566952 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.566977 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.567011 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.567035 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.588382 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184" exitCode=0 Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.588540 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.596922 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.613582 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.637245 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.655815 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.670287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.670358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.670377 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.670405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.670429 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.678268 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.700929 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.720915 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.741777 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.768191 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.781728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.781775 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.781784 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.781803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.781816 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.796270 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.819626 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.863192 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.877991 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.884017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.884054 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.884065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.884085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.884097 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.889764 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.900819 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:54Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.987422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.987471 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.987485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.987506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:54 crc kubenswrapper[4960]: I1201 19:39:54.987522 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:54Z","lastTransitionTime":"2025-12-01T19:39:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.091061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.091168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.091189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.091498 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.091532 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.194163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.194216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.194231 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.194252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.194263 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.297087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.297155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.297167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.297184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.297231 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.324061 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.324259 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:55 crc kubenswrapper[4960]: E1201 19:39:55.324348 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.324374 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:55 crc kubenswrapper[4960]: E1201 19:39:55.324604 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:55 crc kubenswrapper[4960]: E1201 19:39:55.324673 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.343742 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.360317 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.373647 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.398017 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.400858 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.400921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.400936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.400961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.400976 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.417032 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.436075 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.459720 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.479155 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.495244 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.504133 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.504185 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.504195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.504214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.504225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.521350 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.545510 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.563268 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.578047 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.592739 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.607867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.607929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.607946 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.607974 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.607993 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.611296 4960 generic.go:334] "Generic (PLEG): container finished" podID="804735e5-ed5b-4c59-8801-02f492e86d4e" containerID="daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba" exitCode=0 Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.611374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerDied","Data":"daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.632521 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.652591 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.673707 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.692407 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.710694 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.712349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.712386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.712399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.712418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.712434 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.732977 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.753318 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.771198 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.786566 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.804855 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.815912 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.815959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.815973 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.815993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.816022 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.819918 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.833613 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.846111 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.860087 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.918210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.918240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.918251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.918264 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:55 crc kubenswrapper[4960]: I1201 19:39:55.918273 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:55Z","lastTransitionTime":"2025-12-01T19:39:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.021711 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.022034 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.022195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.022272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.022356 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.124841 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.124887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.124897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.124913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.124925 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.227510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.227596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.227624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.227654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.227680 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.330998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.331064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.331083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.331106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.331175 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.434342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.434418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.434437 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.434463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.434486 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.537614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.537685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.537706 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.537735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.537756 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.627854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.628218 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.637161 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" event={"ID":"804735e5-ed5b-4c59-8801-02f492e86d4e","Type":"ContainerStarted","Data":"61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.641419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.641528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.641545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.641570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.641605 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.650803 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.669654 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.673299 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.688107 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.710094 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.731257 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.745103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.745164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.745174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.745189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.745199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.750016 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.774151 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.788639 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.805851 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.819953 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.835907 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.848469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.848545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.848571 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.848604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.848630 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.860602 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.874403 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.906007 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.929062 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.949159 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.950807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.950854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.950874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.950903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.950919 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:56Z","lastTransitionTime":"2025-12-01T19:39:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.967345 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.983207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:56 crc kubenswrapper[4960]: I1201 19:39:56.994271 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:56Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.009571 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.023881 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.035380 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.048974 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.053386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.053423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.053433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.053450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.053461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.073233 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.093573 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.111608 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.131207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.145977 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.156091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.156161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.156178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.156198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.156213 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.259448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.259506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.259523 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.259542 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.259556 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.323813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.323833 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:57 crc kubenswrapper[4960]: E1201 19:39:57.323969 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.323873 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:57 crc kubenswrapper[4960]: E1201 19:39:57.324184 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:57 crc kubenswrapper[4960]: E1201 19:39:57.324241 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.365234 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.365276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.365288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.365305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.365319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.468349 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.468430 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.468450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.468483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.468508 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.571443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.571511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.571528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.571553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.571571 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.641741 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.642822 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.674353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.674408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.674425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.674448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.674462 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.684984 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.709750 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.731614 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.746880 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.770748 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.776860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.776888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.776896 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.776913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.776927 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.786810 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.806255 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.828890 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.847160 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.863365 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.881080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.881240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.881261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.881291 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.881310 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.884788 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.914849 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.931566 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.945796 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.960892 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:57Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.983965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.984020 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.984033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.984052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:57 crc kubenswrapper[4960]: I1201 19:39:57.984067 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:57Z","lastTransitionTime":"2025-12-01T19:39:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.086557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.086620 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.086634 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.086651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.086665 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.189290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.189362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.189374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.189396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.189409 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.292005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.292051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.292062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.292078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.292089 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.357853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.357902 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.357914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.357929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.357942 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.369220 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:58Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.372267 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.372314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.372333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.372352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.372366 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.383270 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:58Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.385769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.385807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.385818 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.385836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.385848 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.398117 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:58Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.401398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.401426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.401435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.401449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.401461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.412377 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:58Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.415467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.415499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.415511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.415527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.415538 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.426730 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:58Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:58 crc kubenswrapper[4960]: E1201 19:39:58.426836 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.428183 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.428228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.428241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.428257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.428273 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.531680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.531764 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.531792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.531823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.531848 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.635342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.635426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.635446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.635471 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.635489 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.644215 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.737754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.737793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.737804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.737817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.737826 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.840230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.840290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.840303 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.840324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.840337 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.943422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.943455 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.943466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.943480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:58 crc kubenswrapper[4960]: I1201 19:39:58.943489 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:58Z","lastTransitionTime":"2025-12-01T19:39:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.047578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.047647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.047666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.047690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.047708 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.151083 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.151145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.151157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.151174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.151187 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.254594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.254674 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.254694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.254721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.254743 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.323777 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.323853 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:39:59 crc kubenswrapper[4960]: E1201 19:39:59.324033 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.324158 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:39:59 crc kubenswrapper[4960]: E1201 19:39:59.324360 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:39:59 crc kubenswrapper[4960]: E1201 19:39:59.324483 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.357730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.357813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.357840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.357895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.357924 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.411268 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.428227 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.452234 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.461527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.461638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.461658 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.461685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.461704 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.473794 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.494466 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.514263 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.531628 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.546209 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.565344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.565416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.565444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.565469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.565489 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.572519 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.590294 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.611000 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.630298 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.642219 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.649779 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/0.log" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.653247 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3" exitCode=1 Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.653307 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.653991 4960 scope.go:117] "RemoveContainer" containerID="f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668660 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.668641 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.687689 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.706507 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.727286 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.749098 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.772452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.772778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.772899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.772986 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.773075 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.775968 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.793277 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.808673 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.838579 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"message\\\":\\\" 6301 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:39:58.886786 6301 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:39:58.886793 6301 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886802 6301 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:39:58.886820 6301 factory.go:656] Stopping watch factory\\\\nI1201 19:39:58.886838 6301 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 19:39:58.886840 6301 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:39:58.886853 6301 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:39:58.886895 6301 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886897 6301 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886973 6301 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887150 6301 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887315 6301 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.853512 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.869202 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.875998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.876037 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.876048 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.876065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.876077 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.883412 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.897443 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.920138 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.937485 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.957460 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:39:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.979534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.979625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.979655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.979688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:39:59 crc kubenswrapper[4960]: I1201 19:39:59.979710 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:39:59Z","lastTransitionTime":"2025-12-01T19:39:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.082917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.083051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.083105 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.083173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.083193 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.186232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.186296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.186313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.186338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.186355 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.289479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.289560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.289586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.289623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.289650 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.368800 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.368938 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.368980 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369157 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369107 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:40:16.369055762 +0000 UTC m=+51.656547481 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369181 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369238 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369271 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369284 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369305 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.369302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369350 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:16.36933441 +0000 UTC m=+51.656826109 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.369401 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369429 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369476 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:16.369462544 +0000 UTC m=+51.656954253 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369495 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369503 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:16.369491025 +0000 UTC m=+51.656982724 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: E1201 19:40:00.369560 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:16.369538287 +0000 UTC m=+51.657029966 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.392148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.392277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.392328 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.392354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.392376 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.495422 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.495812 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.495823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.495839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.495851 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.598887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.598969 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.598988 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.599012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.599028 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.659795 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/0.log" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.663909 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.664054 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.685071 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.697914 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.702382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.702442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.702469 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.702503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.702546 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.709100 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.723421 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.739241 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.750267 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.805595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.805639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.805653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.805671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.805685 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.818214 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.842046 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.854552 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.871273 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.887557 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.899850 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.908409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.908459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.908476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.908497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.908510 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:00Z","lastTransitionTime":"2025-12-01T19:40:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.912267 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:00 crc kubenswrapper[4960]: I1201 19:40:00.929029 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"message\\\":\\\" 6301 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:39:58.886786 6301 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:39:58.886793 6301 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886802 6301 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:39:58.886820 6301 factory.go:656] Stopping watch factory\\\\nI1201 19:39:58.886838 6301 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 19:39:58.886840 6301 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:39:58.886853 6301 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:39:58.886895 6301 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886897 6301 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886973 6301 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887150 6301 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887315 6301 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.011493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.011548 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.011561 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.011580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.011594 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.115176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.115246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.115265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.115291 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.115311 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.218436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.218500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.218518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.218543 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.218560 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.322519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.322578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.322602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.322627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.322644 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.323352 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.323492 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:01 crc kubenswrapper[4960]: E1201 19:40:01.323666 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.323731 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:01 crc kubenswrapper[4960]: E1201 19:40:01.323861 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:01 crc kubenswrapper[4960]: E1201 19:40:01.323972 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.425515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.425578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.425597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.425628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.425650 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.529084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.529198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.529224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.529255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.529280 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.632792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.632848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.632862 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.632881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.632895 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.669725 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/1.log" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.670393 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/0.log" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.673463 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" exitCode=1 Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.673513 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.673593 4960 scope.go:117] "RemoveContainer" containerID="f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.674407 4960 scope.go:117] "RemoveContainer" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" Dec 01 19:40:01 crc kubenswrapper[4960]: E1201 19:40:01.674658 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.694211 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.712930 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.727084 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.737825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.737874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.737887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.737907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.737916 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.743950 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.758396 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.772273 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.788769 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.806523 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"message\\\":\\\" 6301 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:39:58.886786 6301 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:39:58.886793 6301 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886802 6301 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:39:58.886820 6301 factory.go:656] Stopping watch factory\\\\nI1201 19:39:58.886838 6301 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 19:39:58.886840 6301 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:39:58.886853 6301 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:39:58.886895 6301 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886897 6301 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886973 6301 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887150 6301 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887315 6301 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.818847 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.830375 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.841984 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.842095 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.842152 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.842183 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.842208 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.842491 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.854999 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.871019 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.884058 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.921490 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt"] Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.922073 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.924566 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.924887 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.941652 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.945130 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.945175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.945187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.945206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.945220 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:01Z","lastTransitionTime":"2025-12-01T19:40:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.959197 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.975228 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:01 crc kubenswrapper[4960]: I1201 19:40:01.996724 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:01Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.013198 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.026255 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.041636 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.047730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.047780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.047794 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.047823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.047837 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.066987 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e3a075b7f549f9d45de4d8ea84bc2a17721a6d2f502055460bd46cc4c83aa3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"message\\\":\\\" 6301 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:39:58.886786 6301 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:39:58.886793 6301 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886802 6301 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:39:58.886820 6301 factory.go:656] Stopping watch factory\\\\nI1201 19:39:58.886838 6301 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1201 19:39:58.886840 6301 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:39:58.886853 6301 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:39:58.886895 6301 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886897 6301 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.886973 6301 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887150 6301 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:39:58.887315 6301 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.082611 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.092100 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.092185 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.092239 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.092263 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwwr\" (UniqueName: \"kubernetes.io/projected/68b26761-90e5-4d91-8176-99e0cd4aa3eb-kube-api-access-7fwwr\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.094956 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.110562 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.125571 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.140985 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.150638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.150882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.151068 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.151311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.151507 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.155552 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.169824 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.193662 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.193935 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.194192 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.194368 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwwr\" (UniqueName: \"kubernetes.io/projected/68b26761-90e5-4d91-8176-99e0cd4aa3eb-kube-api-access-7fwwr\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.195327 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.195472 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68b26761-90e5-4d91-8176-99e0cd4aa3eb-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.204038 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68b26761-90e5-4d91-8176-99e0cd4aa3eb-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.222017 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwwr\" (UniqueName: \"kubernetes.io/projected/68b26761-90e5-4d91-8176-99e0cd4aa3eb-kube-api-access-7fwwr\") pod \"ovnkube-control-plane-749d76644c-pz5dt\" (UID: \"68b26761-90e5-4d91-8176-99e0cd4aa3eb\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.241701 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.254449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.254496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.254514 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.254539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.254557 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: W1201 19:40:02.262738 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68b26761_90e5_4d91_8176_99e0cd4aa3eb.slice/crio-65ce5fbb18ed68a290a3f1aa4e734629305d83371e4bbb07e0645b764d1d652a WatchSource:0}: Error finding container 65ce5fbb18ed68a290a3f1aa4e734629305d83371e4bbb07e0645b764d1d652a: Status 404 returned error can't find the container with id 65ce5fbb18ed68a290a3f1aa4e734629305d83371e4bbb07e0645b764d1d652a Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.357948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.358019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.358033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.358056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.358075 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.461870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.461916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.461927 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.461943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.461958 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.565346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.565400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.565417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.565436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.565449 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.668270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.668321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.668334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.668353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.668368 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.680325 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" event={"ID":"68b26761-90e5-4d91-8176-99e0cd4aa3eb","Type":"ContainerStarted","Data":"0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.680388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" event={"ID":"68b26761-90e5-4d91-8176-99e0cd4aa3eb","Type":"ContainerStarted","Data":"65ce5fbb18ed68a290a3f1aa4e734629305d83371e4bbb07e0645b764d1d652a"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.682816 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/1.log" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.688652 4960 scope.go:117] "RemoveContainer" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" Dec 01 19:40:02 crc kubenswrapper[4960]: E1201 19:40:02.689056 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.709733 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.722343 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.745397 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.761331 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.771413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.771466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.771478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.771499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.771512 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.779651 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.798199 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.814976 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.833220 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.847727 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.865568 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.874998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.875053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.875062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.875079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.875090 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.880841 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.896476 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.919476 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.946863 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.966557 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.978517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.978573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.978584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.978604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:02 crc kubenswrapper[4960]: I1201 19:40:02.978616 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:02Z","lastTransitionTime":"2025-12-01T19:40:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.064695 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jsbh7"] Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.065479 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.065582 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.082106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.082224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.082246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.082274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.082296 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.089381 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.110830 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.130507 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.149048 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.176350 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.185849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.185909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.185923 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.185943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.185959 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.203553 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.206046 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.206340 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqztw\" (UniqueName: \"kubernetes.io/projected/30d9bdd8-31e7-460b-aefb-421c6d306fbd-kube-api-access-nqztw\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.226865 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.256513 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.277911 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.289027 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.289070 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.289082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.289099 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.289149 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.299436 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.307819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqztw\" (UniqueName: \"kubernetes.io/projected/30d9bdd8-31e7-460b-aefb-421c6d306fbd-kube-api-access-nqztw\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.307880 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.308050 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.308140 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:03.808098486 +0000 UTC m=+39.095590155 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.316737 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.323577 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.323692 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.323734 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.323902 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.324148 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.324256 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.330566 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqztw\" (UniqueName: \"kubernetes.io/projected/30d9bdd8-31e7-460b-aefb-421c6d306fbd-kube-api-access-nqztw\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.335111 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.353364 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.372567 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.390159 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.391436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.391488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.391506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.391531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.391545 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.414602 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.495317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.495393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.495417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.495446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.495478 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.598305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.598380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.598406 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.598439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.598463 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.694293 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" event={"ID":"68b26761-90e5-4d91-8176-99e0cd4aa3eb","Type":"ContainerStarted","Data":"03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.701543 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.701599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.701618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.701644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.701664 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.713390 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.731908 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.750913 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.766455 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.782167 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.801329 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.804035 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.804090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.804109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.804171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.804190 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.815317 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.815528 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:03 crc kubenswrapper[4960]: E1201 19:40:03.815603 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:04.815583005 +0000 UTC m=+40.103074674 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.826829 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.839447 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.855707 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.872487 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.887461 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.901844 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.907186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.907233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.907244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.907261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.907272 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:03Z","lastTransitionTime":"2025-12-01T19:40:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.917097 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.936273 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.952515 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:03 crc kubenswrapper[4960]: I1201 19:40:03.965410 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:03Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.011251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.011333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.011353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.011384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.011418 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.114280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.114331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.114347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.114371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.114391 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.217633 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.217680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.217690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.217706 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.217716 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.237611 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.238952 4960 scope.go:117] "RemoveContainer" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" Dec 01 19:40:04 crc kubenswrapper[4960]: E1201 19:40:04.239236 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.323010 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.323093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.323177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.323220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.323262 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.426868 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.426943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.426962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.426990 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.427010 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.530238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.530324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.530344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.530364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.530379 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.633536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.633627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.633644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.633670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.633687 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.736895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.736954 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.736972 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.736995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.737012 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.831168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:04 crc kubenswrapper[4960]: E1201 19:40:04.831469 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:04 crc kubenswrapper[4960]: E1201 19:40:04.831564 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:06.831538553 +0000 UTC m=+42.119030252 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.840667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.840749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.840774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.840994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.841018 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.943977 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.944065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.944086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.944157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:04 crc kubenswrapper[4960]: I1201 19:40:04.944196 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:04Z","lastTransitionTime":"2025-12-01T19:40:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.046597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.046671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.046699 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.046910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.046930 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.149735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.149822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.149847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.149879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.149901 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.253728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.253793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.253812 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.253838 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.253857 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.323634 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.323693 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.323740 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.323837 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:05 crc kubenswrapper[4960]: E1201 19:40:05.323937 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:05 crc kubenswrapper[4960]: E1201 19:40:05.324079 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:05 crc kubenswrapper[4960]: E1201 19:40:05.324243 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:05 crc kubenswrapper[4960]: E1201 19:40:05.324296 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.344012 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.357655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.357716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.357730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.357753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.357768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.359506 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.375915 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.390386 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.405894 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.425969 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.445256 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.460028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.460086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.460098 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.460139 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.460154 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.462317 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.478373 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.492484 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.510290 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.529889 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.545836 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.561043 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.562663 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.562738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.562754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.562777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.562790 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.583939 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.598993 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.665352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.665404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.665416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.665439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.665454 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.769097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.769218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.769246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.769281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.769311 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.871909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.871976 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.871995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.872021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.872038 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.975178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.975271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.975299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.975338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:05 crc kubenswrapper[4960]: I1201 19:40:05.975365 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:05Z","lastTransitionTime":"2025-12-01T19:40:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.078373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.078457 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.078476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.078514 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.078542 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.183287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.183380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.183404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.183438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.183460 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.287200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.287281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.287316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.287350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.287380 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.391843 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.391914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.391926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.391947 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.391960 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.495516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.495578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.495589 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.495616 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.495628 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.598357 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.598417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.598435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.598458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.598475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.701199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.701286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.701311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.701343 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.701367 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.805413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.805481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.805504 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.805535 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.805556 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.856635 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:06 crc kubenswrapper[4960]: E1201 19:40:06.856889 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:06 crc kubenswrapper[4960]: E1201 19:40:06.856977 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:10.856953833 +0000 UTC m=+46.144445542 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.908560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.908636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.908646 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.908665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:06 crc kubenswrapper[4960]: I1201 19:40:06.908679 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:06Z","lastTransitionTime":"2025-12-01T19:40:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.011850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.011967 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.012018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.012051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.012073 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.114931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.114994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.115034 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.115064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.115084 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.218333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.218398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.218417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.218442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.218461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.322366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.322516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.322535 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.322558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.322574 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.323637 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.323678 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:07 crc kubenswrapper[4960]: E1201 19:40:07.323852 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.323932 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:07 crc kubenswrapper[4960]: E1201 19:40:07.324027 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:07 crc kubenswrapper[4960]: E1201 19:40:07.324310 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.324428 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:07 crc kubenswrapper[4960]: E1201 19:40:07.324581 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.427271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.427352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.427373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.427399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.427418 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.530730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.530783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.530800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.530821 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.530838 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.634480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.634560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.634580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.634604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.634621 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.738067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.738178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.738198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.738226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.738244 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.841629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.841719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.841738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.841760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.841776 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.945527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.945591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.945604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.945627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:07 crc kubenswrapper[4960]: I1201 19:40:07.945643 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:07Z","lastTransitionTime":"2025-12-01T19:40:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.048904 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.048961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.048979 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.049003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.049023 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.151960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.152022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.152039 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.152067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.152085 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.255554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.255617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.255625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.255644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.255657 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.359555 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.359607 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.359617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.359635 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.359647 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.462552 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.462625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.462637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.462655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.462670 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.539726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.539815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.539836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.539869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.539888 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.559179 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:08Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.563714 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.563785 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.563806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.563835 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.563863 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.583285 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:08Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.589461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.589518 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.589541 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.589569 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.589590 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.607749 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:08Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.612733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.612811 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.612835 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.612863 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.612886 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.634617 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:08Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.639934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.640099 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.640143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.640169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.640186 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.659696 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:08Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:08 crc kubenswrapper[4960]: E1201 19:40:08.659942 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.662225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.662292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.662311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.662338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.662357 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.766190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.766268 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.766289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.766323 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.766365 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.869611 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.869671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.869687 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.869708 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.869725 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.973397 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.973462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.973478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.973502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:08 crc kubenswrapper[4960]: I1201 19:40:08.973519 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:08Z","lastTransitionTime":"2025-12-01T19:40:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.076062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.076444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.076699 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.076923 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.077151 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.180481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.180556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.180577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.180604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.180624 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.284452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.284540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.284565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.284598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.284621 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.323341 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.323449 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.323478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:09 crc kubenswrapper[4960]: E1201 19:40:09.323795 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.323842 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:09 crc kubenswrapper[4960]: E1201 19:40:09.323985 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:09 crc kubenswrapper[4960]: E1201 19:40:09.324195 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:09 crc kubenswrapper[4960]: E1201 19:40:09.324322 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.387078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.387154 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.387169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.387187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.387198 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.490433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.490499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.490517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.490540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.490564 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.593729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.593782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.593792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.593813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.593825 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.697245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.697326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.697346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.697373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.697396 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.801670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.801705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.801715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.801729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.801740 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.905173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.905246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.905281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.905316 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:09 crc kubenswrapper[4960]: I1201 19:40:09.905337 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:09Z","lastTransitionTime":"2025-12-01T19:40:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.008774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.008826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.008843 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.008866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.008883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.112282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.112347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.112365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.112390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.112410 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.216029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.216449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.216596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.216747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.216891 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.320346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.320689 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.320853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.321009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.321187 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.425850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.425910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.425933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.425958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.425982 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.529792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.529861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.529881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.529907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.529927 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.633324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.633414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.633434 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.633459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.633477 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.736142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.736232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.736252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.736283 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.736305 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.840252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.840795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.841060 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.841310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.841485 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.897374 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:10 crc kubenswrapper[4960]: E1201 19:40:10.897768 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:10 crc kubenswrapper[4960]: E1201 19:40:10.897902 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:18.897870442 +0000 UTC m=+54.185362151 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.945223 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.945288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.945308 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.945332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:10 crc kubenswrapper[4960]: I1201 19:40:10.945351 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:10Z","lastTransitionTime":"2025-12-01T19:40:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.048314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.048378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.048396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.048420 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.048439 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.152187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.152271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.152296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.152328 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.152352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.255704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.255758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.255775 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.255798 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.255817 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.323871 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.323895 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:11 crc kubenswrapper[4960]: E1201 19:40:11.324061 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.324082 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.324187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:11 crc kubenswrapper[4960]: E1201 19:40:11.324211 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:11 crc kubenswrapper[4960]: E1201 19:40:11.324298 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:11 crc kubenswrapper[4960]: E1201 19:40:11.324658 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.358740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.358780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.358794 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.358810 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.358821 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.462715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.462805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.462824 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.462852 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.462870 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.566361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.566448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.566471 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.566509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.566534 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.669579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.669622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.669633 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.669649 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.669659 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.771942 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.771998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.772012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.772033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.772048 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.874875 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.874957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.875030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.875066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.875085 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.978395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.978459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.978477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.978502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:11 crc kubenswrapper[4960]: I1201 19:40:11.978520 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:11Z","lastTransitionTime":"2025-12-01T19:40:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.081926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.082002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.082015 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.082035 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.082050 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.128109 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.145100 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.168804 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.186438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.186492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.186507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.186533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.186549 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.206074 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.228432 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.246453 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.261749 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.279362 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.290329 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.290391 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.290415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.290440 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.290459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.293538 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.306648 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.334893 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.352584 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.366516 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.380160 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.393735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.393783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.393797 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.393816 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.393843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.394824 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.410553 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.433340 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.451396 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:12Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.496942 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.497007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.497022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.497042 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.497058 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.599957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.600013 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.600025 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.600043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.600054 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.703017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.703060 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.703072 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.703090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.703102 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.805869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.805925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.805943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.805967 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.805985 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.908666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.908851 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.908888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.908921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:12 crc kubenswrapper[4960]: I1201 19:40:12.908947 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:12Z","lastTransitionTime":"2025-12-01T19:40:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.012361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.012461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.012485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.012516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.012538 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.116260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.116356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.116385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.116423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.116450 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.220206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.220285 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.220305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.220333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.220353 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323269 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: E1201 19:40:13.323447 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323517 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323506 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323508 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.323623 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:13 crc kubenswrapper[4960]: E1201 19:40:13.323898 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:13 crc kubenswrapper[4960]: E1201 19:40:13.324214 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:13 crc kubenswrapper[4960]: E1201 19:40:13.324365 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.427495 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.427575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.427596 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.427629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.427655 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.530543 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.530640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.530661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.530686 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.530704 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.634408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.634473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.634484 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.634503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.634522 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.737381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.737449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.737467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.737493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.737511 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.841525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.841604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.841627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.841660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.841684 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.944724 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.944790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.944807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.944832 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:13 crc kubenswrapper[4960]: I1201 19:40:13.944851 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:13Z","lastTransitionTime":"2025-12-01T19:40:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.048533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.048599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.048619 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.048643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.048661 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.151715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.151809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.151837 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.151873 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.151895 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.254930 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.255028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.255057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.255092 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.255151 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.357800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.357870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.357900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.357934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.357956 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.460986 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.461038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.461052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.461070 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.461086 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.563997 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.564455 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.564572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.564705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.564833 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.668783 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.669084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.669230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.669320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.669450 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.772375 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.772809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.773061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.773394 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.773631 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.877737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.877812 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.877836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.877870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.877895 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.981971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.982060 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.982078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.982105 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:14 crc kubenswrapper[4960]: I1201 19:40:14.982175 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:14Z","lastTransitionTime":"2025-12-01T19:40:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.085069 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.085152 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.085164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.085182 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.085192 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.189402 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.189476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.189494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.189521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.189544 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.292427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.292506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.292524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.292553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.292572 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.323812 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.323848 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.323959 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:15 crc kubenswrapper[4960]: E1201 19:40:15.324163 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.324291 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:15 crc kubenswrapper[4960]: E1201 19:40:15.324469 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:15 crc kubenswrapper[4960]: E1201 19:40:15.324559 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:15 crc kubenswrapper[4960]: E1201 19:40:15.324686 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.347750 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.367469 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.385862 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.396280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.396338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.396354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.396380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.396397 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.402309 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.416572 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.434043 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.469491 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.484889 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.498676 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.499394 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.499448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.499462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.499481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.499495 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.513688 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.528240 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.541960 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.556706 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.576587 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.591138 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.601688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.601719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.601731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.601749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.601767 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.604753 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.621521 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:15Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.704487 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.704576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.704600 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.704627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.704646 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.808017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.808075 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.808089 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.808132 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.808146 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.911295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.911361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.911385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.911414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:15 crc kubenswrapper[4960]: I1201 19:40:15.911437 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:15Z","lastTransitionTime":"2025-12-01T19:40:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.014796 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.014881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.014906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.014940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.014964 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.118216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.118305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.118332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.118367 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.118394 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.222058 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.222158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.222174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.222201 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.222222 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.325796 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.325859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.325882 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.325907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.325924 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.429830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.429916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.429937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.429969 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.429989 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.450591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.450744 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.450799 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:40:48.450755661 +0000 UTC m=+83.738247370 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.450864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.450904 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.450925 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.450937 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.450998 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:48.450979698 +0000 UTC m=+83.738471367 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.451020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.451045 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451154 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451186 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451199 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451216 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451207 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451241 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:48.451232056 +0000 UTC m=+83.738723715 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451381 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:48.451346679 +0000 UTC m=+83.738838508 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:16 crc kubenswrapper[4960]: E1201 19:40:16.451415 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:40:48.451396651 +0000 UTC m=+83.738888580 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.533858 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.533928 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.533945 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.533970 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.533989 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.638399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.638528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.638546 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.638570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.638620 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.742606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.742675 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.742694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.742718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.742739 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.846735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.846821 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.846846 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.846879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.846901 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.950210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.950384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.950412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.950484 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:16 crc kubenswrapper[4960]: I1201 19:40:16.950502 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:16Z","lastTransitionTime":"2025-12-01T19:40:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.053589 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.053645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.053663 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.053687 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.053708 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.157914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.157993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.158016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.158050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.158077 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.260320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.260400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.260423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.260451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.260474 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.323874 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.323949 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.323887 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.323867 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:17 crc kubenswrapper[4960]: E1201 19:40:17.324041 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:17 crc kubenswrapper[4960]: E1201 19:40:17.324158 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:17 crc kubenswrapper[4960]: E1201 19:40:17.324356 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:17 crc kubenswrapper[4960]: E1201 19:40:17.324489 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.363572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.363627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.363644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.363670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.363688 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.468152 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.468238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.468263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.468297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.468324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.572241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.572344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.572362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.572388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.572408 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.675480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.675549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.675572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.675603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.675627 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.779382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.779456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.779474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.779499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.779517 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.883277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.883341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.883359 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.883384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.883402 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.986325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.986361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.986369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.986383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:17 crc kubenswrapper[4960]: I1201 19:40:17.986394 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:17Z","lastTransitionTime":"2025-12-01T19:40:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.089564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.089642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.089663 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.089691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.089710 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.192926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.192980 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.192997 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.193021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.193041 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.296080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.296202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.296225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.296257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.296278 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.324858 4960 scope.go:117] "RemoveContainer" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.400699 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.401214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.401229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.401274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.401289 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.505540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.505592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.505604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.505624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.505642 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.608421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.608484 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.608493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.608524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.608537 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.713371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.713424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.713443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.713468 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.713488 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.760385 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/1.log" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.764045 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.765079 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.788923 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.810334 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.816725 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.816801 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.816822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.816850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.816871 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.841143 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.864187 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.888164 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.903489 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919324 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919629 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.919651 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.931299 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.959032 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.963919 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.963972 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.963982 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.964000 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.964012 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.975141 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: E1201 19:40:18.977201 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.982336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:18 crc kubenswrapper[4960]: E1201 19:40:18.982714 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:18 crc kubenswrapper[4960]: E1201 19:40:18.982818 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:40:34.982793593 +0000 UTC m=+70.270285292 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.987363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.987438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.987452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.987515 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:18 crc kubenswrapper[4960]: I1201 19:40:18.987531 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:18Z","lastTransitionTime":"2025-12-01T19:40:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:18.999986 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:18Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.006798 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.010922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.010976 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.010987 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.011009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.011022 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.015643 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.024398 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.026845 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.028915 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.028962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.028977 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.028998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.029014 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.043195 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.045511 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.049051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.049140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.049158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.049187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.049205 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.063055 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.064893 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.065046 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.067031 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.067086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.067096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.067133 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.067147 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.081231 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.095204 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.169684 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.169742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.169752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.169769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.169781 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.272668 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.272720 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.272731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.272746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.272758 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.323301 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.323298 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.323340 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.323554 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.323500 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.323790 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.323773 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.323842 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.375467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.375528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.375544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.375567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.375582 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.478436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.478499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.478512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.478534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.478546 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.581189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.581239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.581249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.581270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.581282 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.685302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.685372 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.685390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.685417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.685435 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.772195 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/2.log" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.773486 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/1.log" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.778664 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" exitCode=1 Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.778770 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.778829 4960 scope.go:117] "RemoveContainer" containerID="7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.780103 4960 scope.go:117] "RemoveContainer" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" Dec 01 19:40:19 crc kubenswrapper[4960]: E1201 19:40:19.780623 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.788396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.788467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.788485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.788511 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.788529 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.808235 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.834852 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.854830 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.872816 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.892432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.892475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.892487 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.892506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.892521 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.894567 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.918558 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.945556 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.974913 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.996389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.996447 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.996466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.996493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.996512 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:19Z","lastTransitionTime":"2025-12-01T19:40:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:19 crc kubenswrapper[4960]: I1201 19:40:19.998679 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:19Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.018881 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.056832 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a57084c6c23606a9ed291ce08d377e58f8ad3cc8141d8d114ce39c4378c29d3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"message\\\":\\\" 6428 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1201 19:40:00.913201 6428 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:00.913231 6428 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:00.913254 6428 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:00.913282 6428 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:00.913305 6428 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:00.913310 6428 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:00.913328 6428 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1201 19:40:00.913338 6428 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:00.913346 6428 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:00.913348 6428 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1201 19:40:00.913358 6428 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:00.913366 6428 handler.go:208] Removed *v1.Node event handler 2\\\\nI1201 19:40:00.913371 6428 factory.go:656] Stopping watch factory\\\\nI1201 19:40:00.913381 6428 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1201 19:40:00.913385 6428 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.078098 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100144 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100206 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.100273 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.120714 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.137970 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.153772 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.173382 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.202962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.203030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.203056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.203087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.203142 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.305657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.305727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.305746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.305774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.305802 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.409105 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.409258 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.409279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.409304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.409324 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.512612 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.512675 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.512692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.512715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.512735 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.616724 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.616849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.616926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.616959 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.617021 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.720302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.720365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.720385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.720410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.720456 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.784817 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/2.log" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.791022 4960 scope.go:117] "RemoveContainer" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" Dec 01 19:40:20 crc kubenswrapper[4960]: E1201 19:40:20.791413 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.817758 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.824011 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.824086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.824141 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.824175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.824199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.843857 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.864043 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.885338 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.913159 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.927460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.927513 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.927536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.927562 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.927582 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:20Z","lastTransitionTime":"2025-12-01T19:40:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.935433 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.957541 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:20 crc kubenswrapper[4960]: I1201 19:40:20.984840 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:20Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.005599 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.022328 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.030632 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.030700 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.030722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.030749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.030768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.042596 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.075628 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.095001 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.116557 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.135225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.135307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.135333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.135374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.135398 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.136488 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.154604 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.177235 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:21Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.238609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.238692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.238717 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.238745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.238766 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.323578 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.323729 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:21 crc kubenswrapper[4960]: E1201 19:40:21.323791 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:21 crc kubenswrapper[4960]: E1201 19:40:21.323990 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.324192 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:21 crc kubenswrapper[4960]: E1201 19:40:21.324325 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.324399 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:21 crc kubenswrapper[4960]: E1201 19:40:21.324558 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.341640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.341721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.341748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.341778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.341802 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.445747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.445831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.445857 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.445894 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.445920 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.551915 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.552003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.552062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.552096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.552144 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.655747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.655821 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.655839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.655866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.655888 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.759622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.759693 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.759718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.759749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.759773 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.864026 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.864097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.864148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.864175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.864198 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.967621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.967701 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.967728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.967760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:21 crc kubenswrapper[4960]: I1201 19:40:21.967785 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:21Z","lastTransitionTime":"2025-12-01T19:40:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.071439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.071498 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.071520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.071550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.071574 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.175252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.175323 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.175342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.175366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.175384 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.278357 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.278452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.278478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.278514 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.278542 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.381623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.381706 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.381725 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.381760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.381779 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.485520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.485583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.485599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.485618 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.485632 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.588826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.588889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.588907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.588934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.588951 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.692714 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.692813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.692833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.692872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.692891 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.796527 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.796568 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.796579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.796595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.796610 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.900154 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.900205 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.900224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.900250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:22 crc kubenswrapper[4960]: I1201 19:40:22.900268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:22Z","lastTransitionTime":"2025-12-01T19:40:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.003397 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.003478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.003503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.003535 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.003557 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.106765 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.106805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.106813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.106826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.106834 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.209793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.209860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.209885 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.209916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.209940 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.313059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.313128 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.313140 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.313158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.313169 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.323913 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.323978 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.323934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.323934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:23 crc kubenswrapper[4960]: E1201 19:40:23.324304 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:23 crc kubenswrapper[4960]: E1201 19:40:23.324441 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:23 crc kubenswrapper[4960]: E1201 19:40:23.324577 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:23 crc kubenswrapper[4960]: E1201 19:40:23.324761 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.416913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.416981 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.417001 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.417030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.417051 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.520477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.520566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.520585 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.520637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.520660 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.624503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.624556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.624567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.624580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.624591 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.727064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.727102 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.727151 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.727168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.727178 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.831352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.831847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.832066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.832300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.832467 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.936165 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.936235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.936253 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.936282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:23 crc kubenswrapper[4960]: I1201 19:40:23.936300 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:23Z","lastTransitionTime":"2025-12-01T19:40:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.039099 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.039297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.039325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.039351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.039371 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.142640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.142718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.142738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.142769 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.142791 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.246167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.246245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.246270 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.246300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.246325 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.350175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.350242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.350265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.350293 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.350315 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.453287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.453334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.453344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.453361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.453373 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.556103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.556167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.556176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.556190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.556199 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.659944 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.660034 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.660055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.660080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.660098 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.763512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.763572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.763584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.763602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.763615 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.866993 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.867079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.867097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.867168 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.867188 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.971214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.971296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.971315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.971340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:24 crc kubenswrapper[4960]: I1201 19:40:24.971357 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:24Z","lastTransitionTime":"2025-12-01T19:40:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.074462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.075161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.075182 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.075207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.075225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.178784 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.178860 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.178884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.178912 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.178933 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.284109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.284219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.284245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.284278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.284302 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.323706 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.323802 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.323901 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.323767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:25 crc kubenswrapper[4960]: E1201 19:40:25.324008 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:25 crc kubenswrapper[4960]: E1201 19:40:25.324173 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:25 crc kubenswrapper[4960]: E1201 19:40:25.324436 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:25 crc kubenswrapper[4960]: E1201 19:40:25.324580 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.350963 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.373872 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.388333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.388393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.388411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.388439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.388459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.398929 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.420855 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.445676 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.461509 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.491546 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.492148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.492199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.492219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.492254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.492278 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.507210 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.526405 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.545620 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.562035 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.579368 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.595300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.595363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.595382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.595409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.595428 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.599833 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.621210 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.647372 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.671943 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.689852 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:25Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.698055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.698085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.698095 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.698132 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.698142 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.801077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.801156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.801176 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.801196 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.801209 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.904719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.904826 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.904847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.904874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:25 crc kubenswrapper[4960]: I1201 19:40:25.904891 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:25Z","lastTransitionTime":"2025-12-01T19:40:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.008530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.008602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.008623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.008651 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.008671 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.111628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.111701 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.111721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.111751 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.111768 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.214714 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.214795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.214820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.214846 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.214865 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.318354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.318400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.318417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.318441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.318459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.421579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.421680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.421697 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.421723 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.421741 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.525005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.525103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.525177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.525211 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.525234 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.629070 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.629187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.629214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.629245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.629269 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.733094 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.733216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.733245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.733278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.733296 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.837161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.837241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.837262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.837291 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.837316 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.940625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.940719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.940748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.940781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:26 crc kubenswrapper[4960]: I1201 19:40:26.940802 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:26Z","lastTransitionTime":"2025-12-01T19:40:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.043962 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.044014 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.044029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.044050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.044063 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.147343 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.147412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.147430 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.147456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.147475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.250492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.250573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.250598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.250626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.250644 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.328306 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:27 crc kubenswrapper[4960]: E1201 19:40:27.328506 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.328612 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:27 crc kubenswrapper[4960]: E1201 19:40:27.328701 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.328776 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:27 crc kubenswrapper[4960]: E1201 19:40:27.328863 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.328951 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:27 crc kubenswrapper[4960]: E1201 19:40:27.329049 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.353666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.353730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.353754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.353777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.353796 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.457340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.457416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.457433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.457456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.457474 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.560599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.560673 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.560692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.560716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.560734 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.664180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.664235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.664251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.664276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.664294 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.767477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.767529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.767545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.767567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.767587 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.870956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.871008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.871028 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.871052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.871072 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.974151 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.974200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.974212 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.974228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:27 crc kubenswrapper[4960]: I1201 19:40:27.974239 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:27Z","lastTransitionTime":"2025-12-01T19:40:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.077870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.078259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.078288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.078314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.078331 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.181927 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.182442 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.182667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.182900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.183139 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.286502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.286564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.286575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.286592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.286604 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.389591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.389883 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.390212 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.390453 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.390710 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.493888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.493954 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.493967 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.493985 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.493998 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.598169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.598241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.598263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.598294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.598316 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.701971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.702046 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.702071 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.702104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.702177 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.804669 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.804710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.804718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.804732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.804741 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.907216 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.907283 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.907301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.907327 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:28 crc kubenswrapper[4960]: I1201 19:40:28.907345 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:28Z","lastTransitionTime":"2025-12-01T19:40:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.010998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.011049 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.011058 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.011079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.011089 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.114603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.114657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.114672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.114693 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.114705 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.217797 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.217874 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.217891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.217915 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.217931 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.321396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.321454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.321472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.321496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.321513 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.323811 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.323839 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.323942 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.324040 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.324029 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.324214 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.324402 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.324552 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.339848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.339911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.339930 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.339956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.339984 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.352079 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:29Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.357106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.357312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.357340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.357370 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.357391 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.374991 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:29Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.380606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.380650 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.380661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.380678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.380687 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.392765 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:29Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.397550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.397597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.397612 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.397631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.397644 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.414398 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:29Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.418683 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.418733 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.418758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.418780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.418796 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.432011 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:29Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:29 crc kubenswrapper[4960]: E1201 19:40:29.432208 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.434091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.434155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.434172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.434190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.434204 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.537731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.537790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.537807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.537829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.537847 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.641829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.641880 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.641890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.641910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.641921 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.744150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.744193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.744205 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.744221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.744234 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.847775 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.847822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.847833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.847848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.847860 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.951587 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.951630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.951639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.951654 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:29 crc kubenswrapper[4960]: I1201 19:40:29.951663 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:29Z","lastTransitionTime":"2025-12-01T19:40:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.054638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.054712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.054732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.054759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.054778 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.157753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.157804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.157813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.157829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.157841 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.260156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.260231 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.260277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.260307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.260333 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.362799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.362850 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.362861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.362878 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.362889 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.465730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.465779 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.465789 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.465804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.465814 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.568872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.568916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.568924 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.568938 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.568948 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.672653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.672703 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.672719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.672735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.672748 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.775310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.775389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.775408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.775436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.775454 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.878573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.878617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.878627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.878642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.878653 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.982090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.982180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.982193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.982213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:30 crc kubenswrapper[4960]: I1201 19:40:30.982228 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:30Z","lastTransitionTime":"2025-12-01T19:40:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.085384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.085450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.085474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.085501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.085522 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.188776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.188841 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.188862 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.188893 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.188911 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.292550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.292603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.292614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.292631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.292643 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.324053 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:31 crc kubenswrapper[4960]: E1201 19:40:31.324272 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.324325 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:31 crc kubenswrapper[4960]: E1201 19:40:31.324464 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.324335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:31 crc kubenswrapper[4960]: E1201 19:40:31.324597 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.324507 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:31 crc kubenswrapper[4960]: E1201 19:40:31.324723 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.395106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.395184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.395198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.395220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.395234 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.498372 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.498456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.498480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.498513 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.498537 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.601414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.601478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.601489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.601580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.601596 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.704827 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.704897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.704911 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.704930 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.704945 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.807829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.807893 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.807907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.807920 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.807932 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.911345 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.911435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.911463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.911499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:31 crc kubenswrapper[4960]: I1201 19:40:31.911521 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:31Z","lastTransitionTime":"2025-12-01T19:40:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.014869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.015353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.015478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.015625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.015764 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.118672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.118730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.118748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.118777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.118793 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.222049 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.222369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.222446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.222529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.222599 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.331190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.331424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.331595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.331749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.331812 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.436157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.436214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.436232 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.436257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.436276 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.539288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.539644 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.539719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.539822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.539918 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.644875 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.645409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.645599 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.645778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.646356 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.749961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.750671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.750726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.750767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.750788 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.853253 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.853380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.853399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.853426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.853444 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.957087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.957218 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.957274 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.957304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:32 crc kubenswrapper[4960]: I1201 19:40:32.957321 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:32Z","lastTransitionTime":"2025-12-01T19:40:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.060910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.060988 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.061007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.061032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.061050 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.164175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.164224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.164235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.164251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.164261 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.267221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.267286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.267302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.267327 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.267349 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.323609 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.323682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.323631 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.323653 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:33 crc kubenswrapper[4960]: E1201 19:40:33.323941 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:33 crc kubenswrapper[4960]: E1201 19:40:33.324412 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:33 crc kubenswrapper[4960]: E1201 19:40:33.324635 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.324680 4960 scope.go:117] "RemoveContainer" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" Dec 01 19:40:33 crc kubenswrapper[4960]: E1201 19:40:33.324693 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:33 crc kubenswrapper[4960]: E1201 19:40:33.324861 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.370385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.370423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.370433 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.370447 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.370458 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.474157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.474512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.474585 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.474670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.474743 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.577088 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.577163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.577175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.577194 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.577205 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.680778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.680820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.680829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.680844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.680857 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.783966 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.784009 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.784018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.784032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.784043 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.887250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.887315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.887337 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.887371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.887395 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.990847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.990908 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.990920 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.990940 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:33 crc kubenswrapper[4960]: I1201 19:40:33.990954 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:33Z","lastTransitionTime":"2025-12-01T19:40:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.093578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.093643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.093655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.093676 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.093690 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.200478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.200534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.200554 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.200581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.200600 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.303638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.303688 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.303697 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.303718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.303733 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.407063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.407210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.407235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.407262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.407283 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.510456 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.510528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.510544 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.510572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.510597 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.613016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.613062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.613075 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.613097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.613130 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.715390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.715475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.715495 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.715525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.715546 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.818665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.818716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.818728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.818747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.818760 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.921254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.921308 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.921321 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.921344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:34 crc kubenswrapper[4960]: I1201 19:40:34.921358 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:34Z","lastTransitionTime":"2025-12-01T19:40:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.024296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.024354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.024366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.024386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.024400 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.074935 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.075184 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.075282 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:41:07.075258296 +0000 UTC m=+102.362750165 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.126709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.126767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.126790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.126818 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.126842 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.229448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.229801 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.229904 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.230019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.230148 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.323992 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.324057 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.324136 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.324189 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.324271 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.324472 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.324014 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:35 crc kubenswrapper[4960]: E1201 19:40:35.324652 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.332957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.332985 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.332994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.333008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.333020 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.340380 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.356155 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.375687 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.394758 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.420076 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.435363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.435446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.435463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.435507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.435523 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.442881 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.461823 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.477798 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.500201 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.512728 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.530483 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.537806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.537856 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.537869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.537892 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.537906 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.544216 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.556025 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.568191 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.585418 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.599234 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.615043 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.640773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.640830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.640845 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.640869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.640885 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.746240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.746301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.746320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.746340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.746352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.847795 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/0.log" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.847870 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfba83a1-281b-4f8f-be15-2760b2b9dfce" containerID="e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8" exitCode=1 Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.847918 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerDied","Data":"e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.848434 4960 scope.go:117] "RemoveContainer" containerID="e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.848773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.848808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.848819 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.848835 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.849257 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.868460 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.884841 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.896638 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.914907 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.935929 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.950607 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.954227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.954259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.954271 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.954289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.954300 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:35Z","lastTransitionTime":"2025-12-01T19:40:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.963475 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.979788 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:35 crc kubenswrapper[4960]: I1201 19:40:35.997715 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:35Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.015565 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.029091 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.051160 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.056665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.056731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.056749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.056777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.056793 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.072630 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.087857 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.100590 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.134640 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.149179 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.159351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.159415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.159434 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.159465 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.159485 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.262730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.262815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.262833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.262861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.262883 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.366411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.366482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.366503 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.366533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.366550 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.470243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.470346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.470371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.470403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.470427 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.573361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.573408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.573419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.573435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.573446 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.677448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.677505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.677516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.677538 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.677551 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.781257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.781317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.781336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.781369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.781391 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.854560 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/0.log" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.854657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerStarted","Data":"9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.873881 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.884677 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.884743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.884763 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.884792 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.884812 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.900732 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.919299 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.944186 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.962729 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.987802 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.987873 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.987898 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.987931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.987956 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:36Z","lastTransitionTime":"2025-12-01T19:40:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:36 crc kubenswrapper[4960]: I1201 19:40:36.988632 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:36Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.004904 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.021927 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.037745 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.060429 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.080412 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.091292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.091362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.091391 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.091424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.091448 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.102188 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.122040 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.141385 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.164057 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.187592 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.194075 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.194159 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.194173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.194212 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.194225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.217346 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:37Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.297460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.297530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.297549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.297574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.297592 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.323669 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.323718 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.323733 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:37 crc kubenswrapper[4960]: E1201 19:40:37.323881 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.323897 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:37 crc kubenswrapper[4960]: E1201 19:40:37.324025 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:37 crc kubenswrapper[4960]: E1201 19:40:37.324174 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:37 crc kubenswrapper[4960]: E1201 19:40:37.324270 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.401210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.401285 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.401304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.401330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.401353 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.504736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.504784 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.504795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.504833 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.504847 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.607935 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.607999 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.608016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.608036 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.608051 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.711702 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.711803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.711823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.711849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.711870 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.815062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.815137 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.815149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.815169 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.815182 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.918371 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.918464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.918489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.918524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:37 crc kubenswrapper[4960]: I1201 19:40:37.918548 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:37Z","lastTransitionTime":"2025-12-01T19:40:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.021499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.021560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.021576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.021602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.021620 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.124432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.124501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.124519 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.124545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.124566 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.228023 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.228085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.228102 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.228162 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.228184 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.331490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.331578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.331597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.331626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.331646 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.435170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.435243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.435261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.435293 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.435316 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.539666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.539724 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.539743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.539768 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.539789 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.643972 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.644033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.644054 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.644079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.644098 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.747303 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.747384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.747408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.747435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.747453 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.850104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.850174 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.850187 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.850206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.850222 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.952819 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.952854 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.952863 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.952881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:38 crc kubenswrapper[4960]: I1201 19:40:38.952893 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:38Z","lastTransitionTime":"2025-12-01T19:40:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.056703 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.056790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.056814 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.056872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.056895 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.159917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.159973 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.159988 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.160008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.160023 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.262831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.262929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.262957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.262990 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.263019 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.324146 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.324179 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.324236 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.324331 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.324148 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.324541 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.324705 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.324919 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.367380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.367444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.367462 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.367487 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.367506 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.470501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.470571 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.470587 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.470613 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.470631 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.486374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.486458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.486478 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.486505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.486523 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.509020 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.514642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.514724 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.514753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.514787 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.514811 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.538154 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.543795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.543866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.543893 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.543924 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.543948 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.565656 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.570982 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.571036 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.571056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.571088 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.571169 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.590659 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.596227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.596341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.596362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.596427 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.596445 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.617073 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:39Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:39 crc kubenswrapper[4960]: E1201 19:40:39.617333 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.619353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.619402 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.619421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.619444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.619461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.723079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.723173 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.723188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.723217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.723238 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.826673 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.826776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.826800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.826839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.826863 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.930638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.930721 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.930744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.930777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:39 crc kubenswrapper[4960]: I1201 19:40:39.930800 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:39Z","lastTransitionTime":"2025-12-01T19:40:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.034335 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.034454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.034476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.034501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.034517 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.137393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.137530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.137561 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.137594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.137620 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.240295 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.240381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.240405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.240438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.240461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.344352 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.344438 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.344461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.344491 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.344520 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.448167 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.448243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.448261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.448284 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.448301 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.551906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.551965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.551987 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.552010 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.552028 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.656014 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.656104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.656181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.656205 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.656225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.760175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.760312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.760340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.760380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.760407 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.864197 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.864266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.864284 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.864311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.864330 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.967737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.967837 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.967861 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.967888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:40 crc kubenswrapper[4960]: I1201 19:40:40.967907 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:40Z","lastTransitionTime":"2025-12-01T19:40:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.071317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.071393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.071418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.071446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.071464 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.174660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.174739 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.174766 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.174799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.174823 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.279011 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.279150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.279172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.279195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.279213 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.324198 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.324265 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.324327 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:41 crc kubenswrapper[4960]: E1201 19:40:41.324416 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.324429 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:41 crc kubenswrapper[4960]: E1201 19:40:41.324586 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:41 crc kubenswrapper[4960]: E1201 19:40:41.324815 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:41 crc kubenswrapper[4960]: E1201 19:40:41.325024 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.382425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.382532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.382560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.382594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.382623 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.485774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.485818 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.485830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.485849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.485862 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.588782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.588843 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.588856 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.588884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.588899 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.692318 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.692396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.692416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.692444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.692467 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.795948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.796007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.796022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.796042 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.796057 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.900177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.900260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.900288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.900322 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:41 crc kubenswrapper[4960]: I1201 19:40:41.900352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:41Z","lastTransitionTime":"2025-12-01T19:40:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.004221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.004306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.004333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.004363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.004387 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.108594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.108681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.108700 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.108727 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.108746 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.211655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.211729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.211749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.211777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.211800 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.315025 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.315108 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.315181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.315214 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.315241 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.419336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.419395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.419407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.419432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.419453 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.522500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.522556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.522572 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.522595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.522610 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.626189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.626277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.626301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.626332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.626353 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.729559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.729649 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.729668 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.729700 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.729720 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.833324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.833401 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.833419 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.833446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.833465 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.937455 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.937536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.937564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.937598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:42 crc kubenswrapper[4960]: I1201 19:40:42.937626 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:42Z","lastTransitionTime":"2025-12-01T19:40:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.041710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.041780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.041795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.041815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.041830 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.145184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.145227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.145237 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.145254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.145266 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.248347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.248426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.248448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.248476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.248495 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.324039 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.324210 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.324584 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:43 crc kubenswrapper[4960]: E1201 19:40:43.324549 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.324653 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:43 crc kubenswrapper[4960]: E1201 19:40:43.324817 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:43 crc kubenswrapper[4960]: E1201 19:40:43.324997 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:43 crc kubenswrapper[4960]: E1201 19:40:43.325414 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.345614 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.351477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.351534 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.351553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.351578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.351601 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.455452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.455508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.455549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.455582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.455603 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.558640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.558736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.558773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.558800 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.558819 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.661289 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.661363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.661384 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.661415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.661435 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.764937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.765017 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.765051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.765085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.765107 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.867788 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.867890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.867921 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.867960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.867987 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.972043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.972156 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.972184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.972213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:43 crc kubenswrapper[4960]: I1201 19:40:43.972235 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:43Z","lastTransitionTime":"2025-12-01T19:40:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.075914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.075979 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.076002 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.076033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.076055 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.180029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.180079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.180097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.180143 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.180157 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.283623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.283692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.283710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.283737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.283752 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.387319 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.387403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.387424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.387451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.387470 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.490774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.490836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.490855 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.490878 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.490898 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.593499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.593549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.593562 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.593579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.593593 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.697107 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.697219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.697238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.697266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.697285 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.800980 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.801051 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.801074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.801107 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.801169 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.904488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.904582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.904604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.904627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:44 crc kubenswrapper[4960]: I1201 19:40:44.904645 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:44Z","lastTransitionTime":"2025-12-01T19:40:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.008300 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.008411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.008432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.008459 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.008477 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.112084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.112198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.112315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.112354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.112379 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.216357 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.216416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.216436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.216463 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.216481 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.319661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.319744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.319764 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.319791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.319811 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.324301 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.324422 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.324422 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.324586 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:45 crc kubenswrapper[4960]: E1201 19:40:45.324588 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:45 crc kubenswrapper[4960]: E1201 19:40:45.324783 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:45 crc kubenswrapper[4960]: E1201 19:40:45.324982 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:45 crc kubenswrapper[4960]: E1201 19:40:45.325351 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.351391 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.373987 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.397967 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.423664 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.423742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.423646 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.423829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.424032 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.424051 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.454782 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.476718 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.500633 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.527055 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.527165 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.527191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.527222 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.527246 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.532032 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.557097 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.576930 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.592984 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.615469 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.630935 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.630996 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.631016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.631044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.631068 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.632976 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.666730 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.688883 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.706000 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.721802 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.734539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.734611 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.734639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.734666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.734693 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.742974 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:45Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.837512 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.837558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.837571 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.837590 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.837605 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.941035 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.941097 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.941158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.941184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:45 crc kubenswrapper[4960]: I1201 19:40:45.941203 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:45Z","lastTransitionTime":"2025-12-01T19:40:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.044749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.044829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.044849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.044877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.044896 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.148719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.148820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.148852 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.148884 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.148905 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.252509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.252580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.252597 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.252620 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.252640 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.324251 4960 scope.go:117] "RemoveContainer" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.357645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.357728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.357760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.357793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.357818 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.460753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.460822 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.460839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.460864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.460885 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.563961 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.564010 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.564025 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.564052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.564064 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.667397 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.667446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.667455 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.667470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.667479 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.770682 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.770718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.770728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.770745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.770759 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.873284 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.873366 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.873391 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.873425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.873451 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.896554 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/2.log" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.899804 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.900504 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.925618 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.947994 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.973017 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.976269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.976306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.976320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.976339 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.976352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:46Z","lastTransitionTime":"2025-12-01T19:40:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:46 crc kubenswrapper[4960]: I1201 19:40:46.991725 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:46Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.007989 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.031284 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.046213 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.061069 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.079972 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.080695 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.080777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.080808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.080838 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.080859 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.105643 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.122046 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.148523 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.173242 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.184003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.184048 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.184059 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.184078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.184090 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.186600 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.201300 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.219553 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.239319 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.257085 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:47Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.287999 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.288050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.288065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.288085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.288099 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.323265 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.323407 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.323404 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:47 crc kubenswrapper[4960]: E1201 19:40:47.323443 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.323506 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:47 crc kubenswrapper[4960]: E1201 19:40:47.323868 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:47 crc kubenswrapper[4960]: E1201 19:40:47.324142 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:47 crc kubenswrapper[4960]: E1201 19:40:47.324248 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.340143 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.391329 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.391411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.391426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.391450 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.391466 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.494481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.494555 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.494573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.494602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.494621 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.597925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.597998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.598008 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.598029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.598042 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.701679 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.701759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.701782 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.701813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.701835 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.805332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.805397 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.805417 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.805445 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.805467 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.907834 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.907905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.907928 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.907960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:47 crc kubenswrapper[4960]: I1201 19:40:47.907985 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:47Z","lastTransitionTime":"2025-12-01T19:40:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.011537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.011605 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.011623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.011647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.011666 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.114448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.114509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.114526 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.114551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.114571 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.218199 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.218250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.218266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.218288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.218306 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.321056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.321157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.321178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.321206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.321225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.424142 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.424210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.424230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.424255 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.424279 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.527496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.527575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.527593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.527617 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.527636 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.531186 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.531359 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531406 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:52.531373584 +0000 UTC m=+147.818865293 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.531458 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.531558 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.531594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531563 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531657 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531677 4960 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531725 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531751 4960 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531771 4960 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531789 4960 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531753 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-01 19:41:52.531730195 +0000 UTC m=+147.819221904 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531832 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:41:52.531817568 +0000 UTC m=+147.819309277 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531616 4960 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531892 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-01 19:41:52.53187984 +0000 UTC m=+147.819371539 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.531934 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-01 19:41:52.531911531 +0000 UTC m=+147.819403210 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.630296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.630363 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.630386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.630415 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.630434 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.734146 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.734204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.734220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.734240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.734253 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.836692 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.836754 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.836774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.836799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.836820 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.911842 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/3.log" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.912915 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/2.log" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.917732 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" exitCode=1 Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.917786 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.917834 4960 scope.go:117] "RemoveContainer" containerID="7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.919284 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:40:48 crc kubenswrapper[4960]: E1201 19:40:48.919660 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.939770 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.939852 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.939879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.939910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.939937 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:48Z","lastTransitionTime":"2025-12-01T19:40:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.940787 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.953741 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.971241 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:48 crc kubenswrapper[4960]: I1201 19:40:48.996295 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:48Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254699 7040 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254790 7040 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.255561 7040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:47.255623 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:47.255633 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:47.255666 7040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:47.255687 7040 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:47.255696 7040 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:47.255713 7040 factory.go:656] Stopping watch factory\\\\nI1201 19:40:47.255727 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:47.255735 7040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:47.255752 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:47.255760 7040 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:47.255773 7040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:48Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.010495 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.043539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.043625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.043643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.043667 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.043714 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.045686 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.069049 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.086712 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.102288 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.118610 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.133711 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f16902-aea4-4ae8-ae6f-281b1f68eb6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36c9bc6acdea251df7b1885d7643ce119ef3535f7ba3d17986090471d91f1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.146881 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.146917 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.146931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.146949 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.146962 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.151754 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.171421 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.187224 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.207467 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.233318 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.250886 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.250937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.250948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.250965 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.250976 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.254490 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.276035 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.299282 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.323896 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.323934 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.323909 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.324082 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.324136 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.324449 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.324412 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.324589 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.354454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.354530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.354553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.354579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.354601 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.457306 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.457382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.457405 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.457435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.457461 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.560423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.560466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.560482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.560507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.560528 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.663273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.663347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.663369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.663398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.663420 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.732372 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.732424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.732439 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.732523 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.732543 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.755985 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.762068 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.762161 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.762178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.762203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.762217 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.783168 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.788555 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.788609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.788628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.788650 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.788665 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.805891 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.811179 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.811230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.811244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.811262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.811275 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.831201 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.836175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.836213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.836227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.836245 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.836260 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.855603 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:49Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:49 crc kubenswrapper[4960]: E1201 19:40:49.855858 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.858006 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.858067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.858087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.858145 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.858165 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.924999 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/3.log" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.961493 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.961548 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.961568 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.961595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:49 crc kubenswrapper[4960]: I1201 19:40:49.961614 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:49Z","lastTransitionTime":"2025-12-01T19:40:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.064550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.064602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.064619 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.064645 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.064662 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.167612 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.167700 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.167719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.167739 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.167759 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.270944 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.271016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.271041 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.271067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.271091 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.373872 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.373989 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.374007 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.374037 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.374055 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.477106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.477219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.477244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.477275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.477295 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.580075 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.580209 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.580237 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.580273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.580300 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.684315 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.684418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.684452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.684492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.684516 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.787317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.787398 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.787418 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.787444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.787463 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.891067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.891200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.891228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.891262 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.891280 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.994451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.994524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.994548 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.994763 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:50 crc kubenswrapper[4960]: I1201 19:40:50.994781 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:50Z","lastTransitionTime":"2025-12-01T19:40:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.097470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.097549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.097577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.097606 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.097627 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.200491 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.200598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.200615 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.200636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.200656 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.303561 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.303640 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.303660 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.303684 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.303702 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.323194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.323239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.323267 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:51 crc kubenswrapper[4960]: E1201 19:40:51.323407 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.323459 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:51 crc kubenswrapper[4960]: E1201 19:40:51.323640 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:51 crc kubenswrapper[4960]: E1201 19:40:51.323787 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:51 crc kubenswrapper[4960]: E1201 19:40:51.323922 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.406158 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.406219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.406238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.406259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.406276 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.509492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.509552 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.509563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.509578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.509610 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.612421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.612681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.612905 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.612955 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.612981 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.716178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.716235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.716246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.716259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.716268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.819588 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.819663 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.819682 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.819707 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.819726 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.923254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.923310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.923330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.923356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:51 crc kubenswrapper[4960]: I1201 19:40:51.923375 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:51Z","lastTransitionTime":"2025-12-01T19:40:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.026853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.026956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.026987 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.027018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.027046 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.129900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.129957 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.129975 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.130001 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.130019 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.234220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.234505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.234662 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.234805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.234952 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.338728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.338793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.338815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.338847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.338869 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.441670 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.441748 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.441777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.441808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.441834 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.545066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.545150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.545171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.545203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.545220 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.649470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.649930 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.650152 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.650397 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.650600 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.754395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.754481 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.754507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.754536 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.754560 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.857664 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.858159 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.858179 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.858204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.858220 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.960506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.960584 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.960608 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.960638 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:52 crc kubenswrapper[4960]: I1201 19:40:52.960659 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:52Z","lastTransitionTime":"2025-12-01T19:40:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.064623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.064722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.064749 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.064778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.064797 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.168178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.168236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.168256 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.168281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.168303 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.271991 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.272046 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.272065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.272090 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.272106 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.323523 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.323603 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.323674 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.323910 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:53 crc kubenswrapper[4960]: E1201 19:40:53.324719 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:53 crc kubenswrapper[4960]: E1201 19:40:53.324895 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:53 crc kubenswrapper[4960]: E1201 19:40:53.324957 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:53 crc kubenswrapper[4960]: E1201 19:40:53.324845 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.378254 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.378319 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.378337 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.378364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.378382 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.481776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.481845 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.481863 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.481891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.481910 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.585082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.585230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.585251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.585280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.585299 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.688621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.688705 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.688722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.688746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.688765 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.792806 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.793248 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.793423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.793611 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.793765 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.897781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.897873 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.897899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.897936 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:53 crc kubenswrapper[4960]: I1201 19:40:53.897959 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:53Z","lastTransitionTime":"2025-12-01T19:40:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.001551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.001622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.001641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.001668 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.001688 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.104794 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.104853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.104871 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.104895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.104913 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.208573 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.208633 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.208652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.208676 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.208696 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.312460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.312522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.312540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.312564 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.312580 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.415460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.415530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.415549 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.415578 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.415596 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.519440 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.519531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.519557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.519592 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.519617 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.623029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.623157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.623184 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.623221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.623247 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.726771 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.726840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.726864 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.726898 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.726919 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.830488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.830550 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.830563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.830583 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.830598 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.933219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.933260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.933269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.933282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:54 crc kubenswrapper[4960]: I1201 19:40:54.933293 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:54Z","lastTransitionTime":"2025-12-01T19:40:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.036542 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.036603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.036621 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.036646 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.036666 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.141047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.141136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.141155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.141175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.141189 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.244044 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.244103 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.244155 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.244178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.244195 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.324065 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.324196 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:55 crc kubenswrapper[4960]: E1201 19:40:55.324280 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.324294 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.324320 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:55 crc kubenswrapper[4960]: E1201 19:40:55.324595 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:55 crc kubenswrapper[4960]: E1201 19:40:55.324916 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:55 crc kubenswrapper[4960]: E1201 19:40:55.325024 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347182 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347362 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347404 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.347424 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.372720 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.393245 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f16902-aea4-4ae8-ae6f-281b1f68eb6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36c9bc6acdea251df7b1885d7643ce119ef3535f7ba3d17986090471d91f1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.423613 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.449533 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.449634 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.449652 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.449737 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.449795 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.458693 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.483207 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.509516 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.531339 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.549890 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.552436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.552509 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.552529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.552556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.552575 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.570961 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.598963 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c3e0016c463476a130ed802064485b0c5976ce298e37c219214c249fd2011db\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:19Z\\\",\\\"message\\\":\\\"ddresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:48Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254699 7040 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254790 7040 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.255561 7040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:47.255623 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:47.255633 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:47.255666 7040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:47.255687 7040 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:47.255696 7040 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:47.255713 7040 factory.go:656] Stopping watch factory\\\\nI1201 19:40:47.255727 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:47.255735 7040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:47.255752 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:47.255760 7040 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:47.255773 7040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.616711 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.640933 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.655973 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.656056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.656079 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.656102 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.656162 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.665893 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.686013 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.703430 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.721972 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.755873 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.759567 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.759616 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.759633 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.759657 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.759674 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.778499 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:55Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.861989 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.862047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.862065 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.862088 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.862105 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.965556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.965614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.965632 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.965655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:55 crc kubenswrapper[4960]: I1201 19:40:55.965673 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:55Z","lastTransitionTime":"2025-12-01T19:40:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.069073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.069189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.069213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.069241 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.069295 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.172018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.172076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.172093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.172148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.172167 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.275538 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.275615 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.275631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.275656 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.275675 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.379898 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.379976 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.379994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.380023 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.380043 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.483809 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.483879 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.483899 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.483924 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.483944 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.587062 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.587162 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.587181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.587204 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.587221 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.689732 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.689820 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.689839 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.689867 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.689888 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.793227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.793317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.793334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.793360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.793380 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.897242 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.897332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.897368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.897399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:56 crc kubenswrapper[4960]: I1201 19:40:56.897422 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:56Z","lastTransitionTime":"2025-12-01T19:40:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.000712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.000763 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.000776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.000795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.000808 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.104043 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.104147 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.104170 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.104194 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.104216 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.208998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.209078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.209101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.209172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.209201 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.312626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.312699 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.312758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.312793 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.313007 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.324231 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.324309 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.324234 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:57 crc kubenswrapper[4960]: E1201 19:40:57.324420 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.324783 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:57 crc kubenswrapper[4960]: E1201 19:40:57.324877 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:57 crc kubenswrapper[4960]: E1201 19:40:57.325045 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:57 crc kubenswrapper[4960]: E1201 19:40:57.325213 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.416958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.417027 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.417045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.417074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.417096 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.521244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.521332 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.521355 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.521387 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.521408 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.625033 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.625101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.625149 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.625175 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.625195 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.728910 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.728971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.728988 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.729015 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.729033 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.832711 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.832796 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.832817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.832847 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.832868 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.936076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.936220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.936244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.936282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:57 crc kubenswrapper[4960]: I1201 19:40:57.936304 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:57Z","lastTransitionTime":"2025-12-01T19:40:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.039510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.039900 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.040107 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.040304 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.040447 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.144815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.145472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.145691 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.145937 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.146212 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.250647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.251106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.251389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.251588 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.251813 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.355529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.355817 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.356040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.356249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.356414 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.460100 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.460225 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.460275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.460313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.460367 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.564163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.564250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.564269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.564293 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.564346 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.668077 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.668190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.668213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.668243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.668267 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.771517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.771591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.771609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.771635 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.771652 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.875563 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.875647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.875671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.875704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.875727 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.979070 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.979181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.979208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.979239 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:58 crc kubenswrapper[4960]: I1201 19:40:58.979263 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:58Z","lastTransitionTime":"2025-12-01T19:40:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.082849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.082912 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.082929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.082949 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.082967 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.186148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.186266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.186284 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.186314 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.186371 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.289452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.289559 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.289576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.289604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.289626 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.323811 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.323842 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.324030 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.324203 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.324272 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.324385 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.324523 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.324623 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.392804 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.392891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.392915 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.392949 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.392978 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.495983 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.496049 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.496063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.496081 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.496094 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.599087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.599180 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.599193 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.599212 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.599225 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.707593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.707719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.707735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.707753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.707763 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.811350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.811428 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.811448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.811474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.811495 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.930350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.930424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.930444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.930473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.930490 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.937778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.937845 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.937869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.937895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.937916 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.960402 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.966888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.966976 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.966994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.967021 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.967039 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:40:59 crc kubenswrapper[4960]: E1201 19:40:59.989968 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:40:59Z is after 2025-08-24T17:21:41Z" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.995472 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.995516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.995528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.995547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:40:59 crc kubenswrapper[4960]: I1201 19:40:59.995561 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:40:59Z","lastTransitionTime":"2025-12-01T19:40:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: E1201 19:41:00.015927 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.021236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.021309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.021328 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.021354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.021370 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: E1201 19:41:00.042557 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.047694 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.047730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.047741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.047758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.047769 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: E1201 19:41:00.064784 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:00Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:00 crc kubenswrapper[4960]: E1201 19:41:00.065558 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.067897 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.068185 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.068380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.068537 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.068747 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.172764 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.173359 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.173591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.173853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.174290 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.278672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.278752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.278766 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.278788 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.278802 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.383273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.383330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.383340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.383357 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.383371 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.486451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.486490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.486500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.486516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.486528 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.589661 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.589726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.589738 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.589759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.589777 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.693467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.693525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.693538 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.693558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.693574 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.796907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.797189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.797210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.797236 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.797253 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.900844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.900926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.900950 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.900979 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:00 crc kubenswrapper[4960]: I1201 19:41:00.901002 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:00Z","lastTransitionTime":"2025-12-01T19:41:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.004052 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.004160 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.004179 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.004206 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.004227 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.106557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.106610 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.106624 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.106647 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.106661 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.210208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.210269 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.210281 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.210301 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.210314 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.314827 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.314883 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.314903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.314929 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.314951 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.323575 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.323642 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:01 crc kubenswrapper[4960]: E1201 19:41:01.323716 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.323575 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.323861 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:01 crc kubenswrapper[4960]: E1201 19:41:01.324118 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:01 crc kubenswrapper[4960]: E1201 19:41:01.324296 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:01 crc kubenswrapper[4960]: E1201 19:41:01.324438 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.418790 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.418851 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.418868 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.418895 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.418913 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.522313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.522423 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.522449 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.522480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.522499 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.625148 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.625217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.625243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.625278 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.625301 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.729054 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.729464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.729659 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.729803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.729956 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.833233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.833325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.833344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.833364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.833379 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.936263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.936337 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.936360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.936388 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:01 crc kubenswrapper[4960]: I1201 19:41:01.936409 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:01Z","lastTransitionTime":"2025-12-01T19:41:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.039566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.039631 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.039649 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.039677 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.039700 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.143317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.143390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.143408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.143435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.143455 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.246287 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.246364 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.246381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.246406 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.246430 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.325037 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:41:02 crc kubenswrapper[4960]: E1201 19:41:02.325402 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.345709 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.350210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.350273 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.350297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.350325 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.350348 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.367836 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.387340 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.419855 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:48Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254699 7040 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254790 7040 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.255561 7040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:47.255623 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:47.255633 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:47.255666 7040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:47.255687 7040 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:47.255696 7040 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:47.255713 7040 factory.go:656] Stopping watch factory\\\\nI1201 19:40:47.255727 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:47.255735 7040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:47.255752 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:47.255760 7040 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:47.255773 7040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.438912 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.454076 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.454164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.454190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.454221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.454242 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.473217 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.498242 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.516246 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.533607 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.548076 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.557305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.557454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.557477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.557502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.557522 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.566511 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f16902-aea4-4ae8-ae6f-281b1f68eb6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36c9bc6acdea251df7b1885d7643ce119ef3535f7ba3d17986090471d91f1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.589854 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.610865 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.632041 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.650203 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.661312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.661382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.661400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.661429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.661448 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.672125 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.692401 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.713702 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.738663 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:02Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.764529 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.764586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.764598 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.764623 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.764660 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.867601 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.867672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.867690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.867719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.867739 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.970360 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.970464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.970488 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.970516 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:02 crc kubenswrapper[4960]: I1201 19:41:02.970539 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:02Z","lastTransitionTime":"2025-12-01T19:41:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.073482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.073522 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.073531 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.073545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.073556 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.175925 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.176000 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.176013 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.176057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.176072 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.279346 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.279411 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.279429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.279452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.279470 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.323256 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.323439 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.323492 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.323454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:03 crc kubenswrapper[4960]: E1201 19:41:03.323656 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:03 crc kubenswrapper[4960]: E1201 19:41:03.323841 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:03 crc kubenswrapper[4960]: E1201 19:41:03.323984 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:03 crc kubenswrapper[4960]: E1201 19:41:03.324062 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.382914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.383038 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.383068 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.383106 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.383168 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.487643 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.487716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.487736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.487760 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.487779 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.590566 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.590637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.590655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.590681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.590701 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.693470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.693524 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.693539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.693558 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.693570 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.797019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.797074 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.797086 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.797104 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.797145 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.900226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.900307 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.900331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.900358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:03 crc kubenswrapper[4960]: I1201 19:41:03.900377 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:03Z","lastTransitionTime":"2025-12-01T19:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.004029 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.004116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.004213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.004244 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.004268 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.107421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.107494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.107507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.107530 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.107543 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.211690 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.211750 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.211767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.211791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.211808 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.315395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.315467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.315490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.315521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.315544 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.418373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.418426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.418436 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.418451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.418465 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.521985 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.522067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.522085 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.522109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.522180 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.624730 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.624797 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.624815 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.624836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.624848 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.728259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.728327 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.728344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.728370 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.728396 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.831575 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.831628 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.831665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.831680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.831688 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.934958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.934995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.935004 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.935018 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:04 crc kubenswrapper[4960]: I1201 19:41:04.935028 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:04Z","lastTransitionTime":"2025-12-01T19:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.038251 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.038312 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.038323 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.038340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.038352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.141775 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.141831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.141840 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.141859 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.141869 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.244525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.244613 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.244630 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.244650 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.244664 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.323890 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.324025 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:05 crc kubenswrapper[4960]: E1201 19:41:05.324096 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.324187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.324185 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:05 crc kubenswrapper[4960]: E1201 19:41:05.324311 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:05 crc kubenswrapper[4960]: E1201 19:41:05.324394 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:05 crc kubenswrapper[4960]: E1201 19:41:05.324451 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.341225 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"68b26761-90e5-4d91-8176-99e0cd4aa3eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a1062a739cda2a7fafbb3c2a1b1414b38f6de3b27cc73c38c39d34964e2e7f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03d29ffd9f67e3da1c6655c5250463432c3e61e7e3a162e3f5b4fe1998d1950f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fwwr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pz5dt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.347576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.347627 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.347655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.347678 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.347692 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.371781 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74e982e8-ef34-4be5-8f97-c52c76b679d1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58a14c9bdaa49cd9a07ae1f2802fd85e1d908750bca8af298de08d9d7a8ae46a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f49a10a1b86e3711ed153d8ceb39975361ca1df25b379f1dbdd11d1ee673610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4da2ff32e4089c129278cacad55049899651a68d3243e9ac05dd11cd366b052a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15c8c50da9adf53680f25ee4105652d0063f0a09aac0b9be3709296f0ab017ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db0346d6664e0aa0a32e22b0869d16de638793ca3214e2b58884b27c8f556dd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c38d6b84f9709e335aec67cc168da24a8c89520a0e5a8cfdd96f2c3ef067d8bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63420417e618e40fe96d63e23793c83af5e18c147ec33ee64885e4d861ef27d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b294b7c08b9a3e266db15ee6e9a730081a3821a9222d3c25a0c2482b0955c5b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.395195 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e18b93b9adc5806b8de16a568fd5a4a24317b2c3bbc7b55b06433a860f360815\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.411381 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lgrxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c645cc4e-73b7-4ca1-8ed4-80ef4ffd2471\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10dc8fd27546981c204a4651997b7e2cea41113223f1669b088031cfb20ba90d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-628zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:48Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lgrxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.421924 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-hcrp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"09fbe457-1574-4501-837a-fb6b55fe46b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c19456acaed6ddd83ca83c9293429e14c04e47b860f9a71a99ce13945cf77429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glscm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:51Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-hcrp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.431932 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85f16902-aea4-4ae8-ae6f-281b1f68eb6b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c36c9bc6acdea251df7b1885d7643ce119ef3535f7ba3d17986090471d91f1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4b51fd1947d84df1b2b38ccefab7ed82deeb78157871bbbfe0cf93b8352175f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.449591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.449642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.449659 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.449680 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.449697 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.461676 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4015e4b-ec52-4bf4-a62e-0cc4b6954dba\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-01T19:39:44Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1201 19:39:38.987460 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1201 19:39:38.989232 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2119040333/tls.crt::/tmp/serving-cert-2119040333/tls.key\\\\\\\"\\\\nI1201 19:39:44.487980 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1201 19:39:44.493671 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1201 19:39:44.493705 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1201 19:39:44.493747 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1201 19:39:44.493777 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1201 19:39:44.505771 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1201 19:39:44.505812 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505820 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1201 19:39:44.505826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1201 19:39:44.505830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1201 19:39:44.505835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1201 19:39:44.505839 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1201 19:39:44.505836 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1201 19:39:44.509423 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.499464 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b495e8efc38934c54e7c598b365dc6f53e2e5673511088b09d2b8b364b494c04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://abd0bc7631f082a6c1b0992990ea9bd9f00f60ec3bf97630ae1a8666f482f4ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.515677 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d32fc0bfd7ccedf9fcfda1438976f2b701772e4e25cb42fd8a92b1103d159db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.533526 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-snl4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"804735e5-ed5b-4c59-8801-02f492e86d4e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61c05c11e7893f908d5ba1b9fa6bf79e62eef920ed927c2c69558a52704b75e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c180b27075390409b4bb99cfd629646eb391b95e68517dad7e702f7be774b84\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ff80fd15768497427eef25129581d46f4d1cdf5c19300f6b07ba30d010805fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d28fce774794a9b92b7fe87943816171213ba55e6f09fa65d97c3bd883eab861\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1caa8618134a6c1afc3c0988355e6015a1c04205eee7fd5e712c1ee84e520f2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://16bebd4971578ab5f71d96d8e96175a455a31249e7c78c8d37afccb4700d8184\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://daf116fb7a827f9333fc28592dfbde3338ae2238506d2636069fc0adf6ca06ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gmtp8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-snl4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.552019 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d788de2e-494c-4730-8cc2-e4cbd8bd397c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e015df5f217b4b5d2ae10bf4b110cbcd1b4d7a327e3840223ba8fa759c4e7a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b572c026c38470a0f6f63fe3cb8756b5a347d0603fbd02259dee2a3beeeda0b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://41a82ccdab70f321beb413b183e5491c3d403b93f633fb425f233b9bf5b314b7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.553094 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.553164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.553178 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.553198 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.553211 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.570115 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f4e5daa2-c655-4572-be66-953a2e358262\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dea8419c63ba0b1ca220de67324ab0d5c97c197ce97759b09b54d59fa593a524\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aef769d30e32b0c619fd781c02bbc593505a233ba60c6d3d5bbab03312aba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d87044f4ffd4662f22b07e766913fcd2eac7a40542b820394d8cbb153ef1c561\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aec258102eaa13bbb31a12d1e6e298db07d865aa8334f2df2461d98dc0393a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.586577 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.605011 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gns76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfba83a1-281b-4f8f-be15-2760b2b9dfce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:35Z\\\",\\\"message\\\":\\\"2025-12-01T19:39:50+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987\\\\n2025-12-01T19:39:50+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_161f9465-38c3-4549-99b6-1ffbe4ba4987 to /host/opt/cni/bin/\\\\n2025-12-01T19:39:50Z [verbose] multus-daemon started\\\\n2025-12-01T19:39:50Z [verbose] Readiness Indicator file check\\\\n2025-12-01T19:40:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:40:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sk8tc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gns76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.624178 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"30d9bdd8-31e7-460b-aefb-421c6d306fbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:40:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqztw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:40:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jsbh7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.642042 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.656353 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.656395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.656407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.656424 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.656438 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.660740 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.677968 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a961900d31e3a5891e95cec251ad2e1014c357d6f798935f885f1748b9e3cbe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-crklw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ct7db\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.702245 4960 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9203b1-b24d-40f4-84e1-180293ad742b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-01T19:40:48Z\\\",\\\"message\\\":\\\"0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254699 7040 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.254790 7040 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1201 19:40:47.255561 7040 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1201 19:40:47.255623 7040 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1201 19:40:47.255633 7040 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1201 19:40:47.255666 7040 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1201 19:40:47.255687 7040 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1201 19:40:47.255696 7040 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1201 19:40:47.255713 7040 factory.go:656] Stopping watch factory\\\\nI1201 19:40:47.255727 7040 ovnkube.go:599] Stopped ovnkube\\\\nI1201 19:40:47.255735 7040 handler.go:208] Removed *v1.Node event handler 7\\\\nI1201 19:40:47.255752 7040 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1201 19:40:47.255760 7040 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1201 19:40:47.255773 7040 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-01T19:40:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-01T19:39:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-01T19:39:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-01T19:39:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5rbhw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-01T19:39:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-fdntq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:05Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.759372 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.759425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.759434 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.759452 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.759464 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.862413 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.862470 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.862483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.862502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.862515 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.965716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.965770 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.965785 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.965805 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:05 crc kubenswrapper[4960]: I1201 19:41:05.965818 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:05Z","lastTransitionTime":"2025-12-01T19:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.068922 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.068979 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.068994 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.069019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.069036 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.172351 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.172426 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.172446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.172473 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.172491 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.276101 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.276230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.276261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.276288 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.276307 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.380246 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.380313 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.380330 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.380354 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.380371 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.483257 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.483329 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.483348 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.483374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.483392 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.585848 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.585903 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.585913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.585931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.585942 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.689432 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.689501 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.689520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.689545 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.689562 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.792774 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.792836 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.792858 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.792887 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.792908 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.896302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.896368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.896387 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.896410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.896427 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.998853 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.998920 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.998932 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.998952 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:06 crc kubenswrapper[4960]: I1201 19:41:06.998966 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:06Z","lastTransitionTime":"2025-12-01T19:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.101711 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.101784 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.101803 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.101825 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.101843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.163952 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.164216 4960 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.164334 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs podName:30d9bdd8-31e7-460b-aefb-421c6d306fbd nodeName:}" failed. No retries permitted until 2025-12-01 19:42:11.164302908 +0000 UTC m=+166.451794617 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs") pod "network-metrics-daemon-jsbh7" (UID: "30d9bdd8-31e7-460b-aefb-421c6d306fbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.204092 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.204192 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.204221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.204248 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.204269 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.307265 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.307333 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.307356 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.307380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.307397 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.323461 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.323575 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.323759 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.323748 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.323782 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.324417 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.324016 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:07 crc kubenswrapper[4960]: E1201 19:41:07.324897 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.412342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.412460 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.412486 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.412521 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.412556 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.515365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.515919 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.516084 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.516299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.516484 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.619553 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.619869 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.619934 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.620056 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.620152 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.723061 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.723120 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.723163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.723188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.723207 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.826889 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.826955 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.826974 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.827003 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.827022 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.930622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.930710 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.930735 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.930768 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:07 crc kubenswrapper[4960]: I1201 19:41:07.930794 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:07Z","lastTransitionTime":"2025-12-01T19:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.033666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.033715 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.033726 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.033741 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.033752 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.137407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.137632 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.137655 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.137683 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.137702 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.241023 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.241078 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.241095 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.241164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.241194 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.344944 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.344990 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.345001 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.345019 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.345031 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.448718 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.448802 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.448823 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.448849 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.448871 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.552807 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.552907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.552926 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.552954 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.552974 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.655998 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.656057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.656080 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.656116 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.656187 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.759745 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.759814 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.759831 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.759858 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.759881 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.863428 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.863505 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.863528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.863557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.863583 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.967499 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.967561 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.967579 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.967604 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:08 crc kubenswrapper[4960]: I1201 19:41:08.967621 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:08Z","lastTransitionTime":"2025-12-01T19:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.070365 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.070443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.070464 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.070490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.070510 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.173506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.173569 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.173586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.173610 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.173630 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.276451 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.276551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.276577 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.276609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.276629 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.323837 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:09 crc kubenswrapper[4960]: E1201 19:41:09.324107 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.324421 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:09 crc kubenswrapper[4960]: E1201 19:41:09.324549 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.324603 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:09 crc kubenswrapper[4960]: E1201 19:41:09.324745 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.324762 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:09 crc kubenswrapper[4960]: E1201 19:41:09.324904 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.380217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.380282 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.380299 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.380324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.380342 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.484136 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.484190 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.484202 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.484220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.484232 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.587684 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.587740 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.587751 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.587773 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.587787 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.690918 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.690963 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.690974 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.690989 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.691001 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.793482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.793552 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.793576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.793600 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.793623 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.896324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.896361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.896369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.896382 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.896391 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.999496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.999551 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.999560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.999574 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:09 crc kubenswrapper[4960]: I1201 19:41:09.999583 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:09Z","lastTransitionTime":"2025-12-01T19:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.104150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.104240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.104266 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.104297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.104319 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.207409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.207466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.207483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.207508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.207526 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.310014 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.310073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.310091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.310115 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.310150 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.376828 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.376877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.376891 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.376909 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.376923 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.393532 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.399082 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.399181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.399200 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.399229 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.399247 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.416025 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.421958 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.422016 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.422040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.422067 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.422086 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.441757 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.447659 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.447722 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.447744 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.447767 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.447786 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.460659 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.465497 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.465565 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.465585 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.465613 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.465637 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.485324 4960 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-01T19:41:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c6e64e4-31f4-48d6-b943-ffee08ff2cb1\\\",\\\"systemUUID\\\":\\\"af0d86f2-b38c-4a68-99f8-21f571d22202\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-01T19:41:10Z is after 2025-08-24T17:21:41Z" Dec 01 19:41:10 crc kubenswrapper[4960]: E1201 19:41:10.485571 4960 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.487625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.487681 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.487704 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.487728 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.487748 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.592050 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.592163 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.592186 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.592217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.592236 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.695906 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.695980 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.695995 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.696022 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.696038 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.800203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.800276 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.800297 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.800326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.800358 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.904483 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.904541 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.904557 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.904582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:10 crc kubenswrapper[4960]: I1201 19:41:10.904601 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:10Z","lastTransitionTime":"2025-12-01T19:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.007982 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.008045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.008063 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.008087 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.008107 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.111171 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.111217 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.111228 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.111247 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.111259 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.215486 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.215562 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.215581 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.215608 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.215627 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.319386 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.319443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.319461 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.319487 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.319506 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.323750 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.323799 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.323799 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:11 crc kubenswrapper[4960]: E1201 19:41:11.323914 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.324181 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:11 crc kubenswrapper[4960]: E1201 19:41:11.324294 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:11 crc kubenswrapper[4960]: E1201 19:41:11.324531 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:11 crc kubenswrapper[4960]: E1201 19:41:11.324776 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.422975 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.423046 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.423064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.423093 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.423116 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.526601 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.526685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.526712 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.526743 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.526766 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.630970 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.631024 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.631040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.631064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.631085 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.734311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.734409 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.734429 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.734455 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.734475 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.838496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.838603 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.838622 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.838648 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.838666 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.941385 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.941507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.941560 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.941586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:11 crc kubenswrapper[4960]: I1201 19:41:11.941641 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:11Z","lastTransitionTime":"2025-12-01T19:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.046390 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.046496 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.046520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.046556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.046593 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.150666 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.150729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.150747 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.150772 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.150790 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.253890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.253943 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.253960 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.254040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.254068 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.356406 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.356485 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.356510 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.356540 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.356559 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.459410 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.459484 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.459502 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.459528 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.459547 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.563109 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.563215 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.563233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.563261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.563283 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.666091 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.666189 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.666208 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.666233 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.666252 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.770069 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.770195 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.770224 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.770259 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.770287 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.873250 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.873323 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.873341 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.873369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.873390 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.977096 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.977191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.977207 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.977231 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:12 crc kubenswrapper[4960]: I1201 19:41:12.977244 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:12Z","lastTransitionTime":"2025-12-01T19:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.081665 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.081736 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.081757 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.081830 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.081885 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.185713 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.185778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.185795 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.185821 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.185843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.289674 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.289742 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.289759 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.289781 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.289797 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.323354 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.323450 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.323476 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.323419 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:13 crc kubenswrapper[4960]: E1201 19:41:13.323629 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:13 crc kubenswrapper[4960]: E1201 19:41:13.323785 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:13 crc kubenswrapper[4960]: E1201 19:41:13.324080 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:13 crc kubenswrapper[4960]: E1201 19:41:13.324615 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.325227 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:41:13 crc kubenswrapper[4960]: E1201 19:41:13.325642 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.393210 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.393261 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.393275 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.393324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.393345 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.497205 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.497283 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.497296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.497317 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.497330 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.600580 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.600637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.600653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.600677 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.600692 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.704334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.704407 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.704425 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.704444 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.704457 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.808243 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.808305 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.808318 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.808340 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.808354 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.913053 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.913172 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.913201 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.913238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:13 crc kubenswrapper[4960]: I1201 19:41:13.913262 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:13Z","lastTransitionTime":"2025-12-01T19:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.016517 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.016591 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.016610 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.016635 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.016655 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.119113 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.119181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.119191 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.119205 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.119214 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.222326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.222408 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.222435 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.222467 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.222493 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.325490 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.325595 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.325614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.325687 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.325714 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.428866 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.428931 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.428948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.428971 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.428990 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.532057 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.532164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.532188 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.532235 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.532256 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.636399 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.636476 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.636507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.636541 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.636563 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.740752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.740890 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.740914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.740975 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.741003 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.844816 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.844888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.844912 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.844948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.844966 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.948851 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.948907 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.948923 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.948948 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:14 crc kubenswrapper[4960]: I1201 19:41:14.948969 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:14Z","lastTransitionTime":"2025-12-01T19:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.052963 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.053045 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.053064 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.053089 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.053110 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.156379 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.156466 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.156489 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.156520 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.156548 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.259933 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.260030 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.260048 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.260073 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.260090 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.323474 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.323621 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:15 crc kubenswrapper[4960]: E1201 19:41:15.323701 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.323717 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:15 crc kubenswrapper[4960]: E1201 19:41:15.323865 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:15 crc kubenswrapper[4960]: E1201 19:41:15.324019 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.323895 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:15 crc kubenswrapper[4960]: E1201 19:41:15.324195 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.362719 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.362808 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.362829 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.362857 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.362878 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.446749 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podStartSLOduration=87.446728546 podStartE2EDuration="1m27.446728546s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.409669123 +0000 UTC m=+110.697160832" watchObservedRunningTime="2025-12-01 19:41:15.446728546 +0000 UTC m=+110.734220225" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.465181 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.465240 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.465252 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.465272 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.465284 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.507546 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=32.507512164 podStartE2EDuration="32.507512164s" podCreationTimestamp="2025-12-01 19:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.506818712 +0000 UTC m=+110.794310391" watchObservedRunningTime="2025-12-01 19:41:15.507512164 +0000 UTC m=+110.795003863" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.547157 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lgrxx" podStartSLOduration=87.547133055 podStartE2EDuration="1m27.547133055s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.546849016 +0000 UTC m=+110.834340675" watchObservedRunningTime="2025-12-01 19:41:15.547133055 +0000 UTC m=+110.834624734" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.558987 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-hcrp4" podStartSLOduration=87.558954666 podStartE2EDuration="1m27.558954666s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.558678068 +0000 UTC m=+110.846169737" watchObservedRunningTime="2025-12-01 19:41:15.558954666 +0000 UTC m=+110.846446335" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.568260 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.568336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.568358 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.568383 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.568398 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.573989 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pz5dt" podStartSLOduration=87.573956014 podStartE2EDuration="1m27.573956014s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.573354016 +0000 UTC m=+110.860845705" watchObservedRunningTime="2025-12-01 19:41:15.573956014 +0000 UTC m=+110.861447723" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.619543 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=91.619515267 podStartE2EDuration="1m31.619515267s" podCreationTimestamp="2025-12-01 19:39:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.619485506 +0000 UTC m=+110.906977215" watchObservedRunningTime="2025-12-01 19:41:15.619515267 +0000 UTC m=+110.907006936" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.620151 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=28.620144536 podStartE2EDuration="28.620144536s" podCreationTimestamp="2025-12-01 19:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.589529851 +0000 UTC m=+110.877021570" watchObservedRunningTime="2025-12-01 19:41:15.620144536 +0000 UTC m=+110.907636205" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.671222 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.671277 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.671290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.671318 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.671330 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.683986 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.683965777 podStartE2EDuration="1m26.683965777s" podCreationTimestamp="2025-12-01 19:39:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.663070018 +0000 UTC m=+110.950561687" watchObservedRunningTime="2025-12-01 19:41:15.683965777 +0000 UTC m=+110.971457446" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.684260 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=63.684254186 podStartE2EDuration="1m3.684254186s" podCreationTimestamp="2025-12-01 19:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.68277564 +0000 UTC m=+110.970267309" watchObservedRunningTime="2025-12-01 19:41:15.684254186 +0000 UTC m=+110.971745855" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.716291 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-gns76" podStartSLOduration=87.716265674 podStartE2EDuration="1m27.716265674s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.715853192 +0000 UTC m=+111.003344861" watchObservedRunningTime="2025-12-01 19:41:15.716265674 +0000 UTC m=+111.003757343" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.735007 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-snl4j" podStartSLOduration=87.734984926 podStartE2EDuration="1m27.734984926s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:15.734281955 +0000 UTC m=+111.021773634" watchObservedRunningTime="2025-12-01 19:41:15.734984926 +0000 UTC m=+111.022476595" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.775514 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.775546 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.775555 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.775570 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.775580 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.878221 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.878286 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.878303 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.878331 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.878351 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.981707 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.981772 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.981791 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.981814 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:15 crc kubenswrapper[4960]: I1201 19:41:15.981833 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:15Z","lastTransitionTime":"2025-12-01T19:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.085213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.085302 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.085324 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.085350 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.085369 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.189894 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.189978 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.190005 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.190040 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.190065 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.293326 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.293393 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.293412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.293437 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.293458 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.397280 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.397357 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.397380 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.397457 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.397476 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.501347 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.501412 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.501431 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.501454 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.501471 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.604507 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.604569 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.604582 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.604636 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.604649 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.708479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.708556 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.708576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.708602 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.708623 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.812047 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.812164 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.812213 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.812238 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.812255 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.915571 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.915614 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.915625 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.915639 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:16 crc kubenswrapper[4960]: I1201 19:41:16.915650 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:16Z","lastTransitionTime":"2025-12-01T19:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.018376 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.018482 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.018506 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.018539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.018560 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.121916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.121985 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.122001 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.122026 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.122043 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.224968 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.225020 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.225041 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.225066 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.225083 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.323516 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:17 crc kubenswrapper[4960]: E1201 19:41:17.323703 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.323744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.323767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:17 crc kubenswrapper[4960]: E1201 19:41:17.323954 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.323767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:17 crc kubenswrapper[4960]: E1201 19:41:17.324197 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:17 crc kubenswrapper[4960]: E1201 19:41:17.324282 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.328400 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.328477 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.328494 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.328514 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.328527 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.432226 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.432294 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.432311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.432336 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.432358 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.535697 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.535761 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.536342 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.536492 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.538352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.642338 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.642403 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.642421 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.642446 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.642466 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.745653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.745714 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.745731 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.745755 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.745774 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.849227 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.849292 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.849310 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.849334 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.849352 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.952368 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.952443 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.952457 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.952479 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:17 crc kubenswrapper[4960]: I1201 19:41:17.952496 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:17Z","lastTransitionTime":"2025-12-01T19:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.055913 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.055979 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.055996 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.056026 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.056047 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.159159 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.159220 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.159230 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.159249 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.159262 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.262685 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.262739 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.262756 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.262777 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.262788 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.366049 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.366157 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.366177 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.366203 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.366227 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.469752 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.469828 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.469844 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.469870 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.469886 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.573637 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.573709 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.573729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.573753 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.573772 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.677474 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.677586 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.677642 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.677672 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.677693 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.781682 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.781758 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.781780 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.781813 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.781838 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.885641 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.885724 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.885746 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.885778 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.885802 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.989219 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.989290 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.989311 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.989339 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:18 crc kubenswrapper[4960]: I1201 19:41:18.989357 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:18Z","lastTransitionTime":"2025-12-01T19:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.092532 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.092609 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.092626 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.092653 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.092674 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.196303 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.196370 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.196389 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.196416 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.196435 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.299662 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.299729 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.299776 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.299799 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.299843 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.323968 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.323967 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.323996 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.324250 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:19 crc kubenswrapper[4960]: E1201 19:41:19.324371 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:19 crc kubenswrapper[4960]: E1201 19:41:19.324658 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:19 crc kubenswrapper[4960]: E1201 19:41:19.324832 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:19 crc kubenswrapper[4960]: E1201 19:41:19.324932 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.403396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.403447 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.403458 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.403475 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.403489 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.507320 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.507395 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.507414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.507441 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.507459 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.610593 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.610669 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.610689 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.610716 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.610736 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.714309 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.714373 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.714396 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.714420 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.714438 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.817296 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.817348 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.817361 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.817378 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.817392 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.919916 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.919990 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.920010 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.920039 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:19 crc kubenswrapper[4960]: I1201 19:41:19.920057 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:19Z","lastTransitionTime":"2025-12-01T19:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.022877 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.022966 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.022985 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.023012 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.023031 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.126263 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.126327 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.126344 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.126369 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.126387 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.232796 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.232888 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.232914 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.232956 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.232978 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.336883 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.336944 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.336964 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.336989 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.337009 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.441150 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.441279 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.441381 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.441414 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.441442 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.544500 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.544576 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.544594 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.544620 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.544640 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.649448 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.649508 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.649525 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.649547 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.649563 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.687374 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.687480 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.687504 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.687539 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.687566 4960 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-01T19:41:20Z","lastTransitionTime":"2025-12-01T19:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.760621 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz"] Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.761250 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.764845 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.764928 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.765015 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.765334 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.830698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.830932 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.831005 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9504f70-9acf-473b-b390-cf13612786ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.831425 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9504f70-9acf-473b-b390-cf13612786ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.831512 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9504f70-9acf-473b-b390-cf13612786ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933214 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9504f70-9acf-473b-b390-cf13612786ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933284 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9504f70-9acf-473b-b390-cf13612786ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933473 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9504f70-9acf-473b-b390-cf13612786ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933563 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.933653 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d9504f70-9acf-473b-b390-cf13612786ff-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.934909 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9504f70-9acf-473b-b390-cf13612786ff-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.942949 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9504f70-9acf-473b-b390-cf13612786ff-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:20 crc kubenswrapper[4960]: I1201 19:41:20.964345 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9504f70-9acf-473b-b390-cf13612786ff-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5zngz\" (UID: \"d9504f70-9acf-473b-b390-cf13612786ff\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:21 crc kubenswrapper[4960]: I1201 19:41:21.082804 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" Dec 01 19:41:21 crc kubenswrapper[4960]: I1201 19:41:21.323525 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:21 crc kubenswrapper[4960]: I1201 19:41:21.323544 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:21 crc kubenswrapper[4960]: I1201 19:41:21.323903 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:21 crc kubenswrapper[4960]: I1201 19:41:21.323958 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:21 crc kubenswrapper[4960]: E1201 19:41:21.324228 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:21 crc kubenswrapper[4960]: E1201 19:41:21.324259 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:21 crc kubenswrapper[4960]: E1201 19:41:21.324398 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:21 crc kubenswrapper[4960]: E1201 19:41:21.324769 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.060182 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" event={"ID":"d9504f70-9acf-473b-b390-cf13612786ff","Type":"ContainerStarted","Data":"6475851713efe458e39d6d325ae366de5825f6643088700a8de359fa8f5f59cf"} Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.060355 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" event={"ID":"d9504f70-9acf-473b-b390-cf13612786ff","Type":"ContainerStarted","Data":"e26b44cde33f91b0fc848ed3e515ca81301c58beca8b79a728e5813fb49fd5d5"} Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.062692 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/1.log" Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.063546 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/0.log" Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.063626 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfba83a1-281b-4f8f-be15-2760b2b9dfce" containerID="9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b" exitCode=1 Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.063661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerDied","Data":"9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b"} Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.063706 4960 scope.go:117] "RemoveContainer" containerID="e7a96f1201e07ef9e095bcce89a9d7f0f5267e3df1c27f8c120aa9eda6a832f8" Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.064421 4960 scope.go:117] "RemoveContainer" containerID="9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b" Dec 01 19:41:22 crc kubenswrapper[4960]: E1201 19:41:22.064909 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-gns76_openshift-multus(cfba83a1-281b-4f8f-be15-2760b2b9dfce)\"" pod="openshift-multus/multus-gns76" podUID="cfba83a1-281b-4f8f-be15-2760b2b9dfce" Dec 01 19:41:22 crc kubenswrapper[4960]: I1201 19:41:22.082594 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5zngz" podStartSLOduration=94.08257109 podStartE2EDuration="1m34.08257109s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:22.08219277 +0000 UTC m=+117.369684499" watchObservedRunningTime="2025-12-01 19:41:22.08257109 +0000 UTC m=+117.370062759" Dec 01 19:41:23 crc kubenswrapper[4960]: I1201 19:41:23.070853 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/1.log" Dec 01 19:41:23 crc kubenswrapper[4960]: I1201 19:41:23.323588 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:23 crc kubenswrapper[4960]: I1201 19:41:23.323641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:23 crc kubenswrapper[4960]: I1201 19:41:23.323680 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:23 crc kubenswrapper[4960]: E1201 19:41:23.323847 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:23 crc kubenswrapper[4960]: I1201 19:41:23.323882 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:23 crc kubenswrapper[4960]: E1201 19:41:23.324012 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:23 crc kubenswrapper[4960]: E1201 19:41:23.324246 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:23 crc kubenswrapper[4960]: E1201 19:41:23.324314 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.313199 4960 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 01 19:41:25 crc kubenswrapper[4960]: I1201 19:41:25.323874 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:25 crc kubenswrapper[4960]: I1201 19:41:25.323957 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:25 crc kubenswrapper[4960]: I1201 19:41:25.323887 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.325357 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:25 crc kubenswrapper[4960]: I1201 19:41:25.325390 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.325570 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.325751 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.325893 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:25 crc kubenswrapper[4960]: E1201 19:41:25.441265 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:41:27 crc kubenswrapper[4960]: I1201 19:41:27.323639 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:27 crc kubenswrapper[4960]: I1201 19:41:27.323639 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:27 crc kubenswrapper[4960]: I1201 19:41:27.323702 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:27 crc kubenswrapper[4960]: E1201 19:41:27.323919 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:27 crc kubenswrapper[4960]: I1201 19:41:27.323792 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:27 crc kubenswrapper[4960]: E1201 19:41:27.324038 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:27 crc kubenswrapper[4960]: E1201 19:41:27.324099 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:27 crc kubenswrapper[4960]: E1201 19:41:27.324635 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:27 crc kubenswrapper[4960]: I1201 19:41:27.324936 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:41:27 crc kubenswrapper[4960]: E1201 19:41:27.325150 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-fdntq_openshift-ovn-kubernetes(8c9203b1-b24d-40f4-84e1-180293ad742b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" Dec 01 19:41:29 crc kubenswrapper[4960]: I1201 19:41:29.323394 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:29 crc kubenswrapper[4960]: I1201 19:41:29.323521 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:29 crc kubenswrapper[4960]: I1201 19:41:29.323532 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:29 crc kubenswrapper[4960]: E1201 19:41:29.323628 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:29 crc kubenswrapper[4960]: E1201 19:41:29.323781 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:29 crc kubenswrapper[4960]: I1201 19:41:29.323912 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:29 crc kubenswrapper[4960]: E1201 19:41:29.323946 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:29 crc kubenswrapper[4960]: E1201 19:41:29.324173 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:30 crc kubenswrapper[4960]: E1201 19:41:30.442847 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:41:31 crc kubenswrapper[4960]: I1201 19:41:31.323573 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:31 crc kubenswrapper[4960]: I1201 19:41:31.323655 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:31 crc kubenswrapper[4960]: I1201 19:41:31.323609 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:31 crc kubenswrapper[4960]: I1201 19:41:31.323584 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:31 crc kubenswrapper[4960]: E1201 19:41:31.323825 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:31 crc kubenswrapper[4960]: E1201 19:41:31.324096 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:31 crc kubenswrapper[4960]: E1201 19:41:31.324277 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:31 crc kubenswrapper[4960]: E1201 19:41:31.324706 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:33 crc kubenswrapper[4960]: I1201 19:41:33.324215 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:33 crc kubenswrapper[4960]: I1201 19:41:33.324239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:33 crc kubenswrapper[4960]: I1201 19:41:33.324241 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:33 crc kubenswrapper[4960]: I1201 19:41:33.324100 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:33 crc kubenswrapper[4960]: E1201 19:41:33.324567 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:33 crc kubenswrapper[4960]: E1201 19:41:33.324770 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:33 crc kubenswrapper[4960]: E1201 19:41:33.325252 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:33 crc kubenswrapper[4960]: E1201 19:41:33.325293 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:33 crc kubenswrapper[4960]: I1201 19:41:33.325360 4960 scope.go:117] "RemoveContainer" containerID="9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b" Dec 01 19:41:34 crc kubenswrapper[4960]: I1201 19:41:34.118011 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/1.log" Dec 01 19:41:34 crc kubenswrapper[4960]: I1201 19:41:34.118085 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerStarted","Data":"323e8982ac01bc4e25a3efd7a695c498c17fb5b84c9453497804840300393a6f"} Dec 01 19:41:35 crc kubenswrapper[4960]: I1201 19:41:35.323505 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:35 crc kubenswrapper[4960]: I1201 19:41:35.323668 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:35 crc kubenswrapper[4960]: I1201 19:41:35.323630 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:35 crc kubenswrapper[4960]: I1201 19:41:35.324222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:35 crc kubenswrapper[4960]: E1201 19:41:35.324758 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:35 crc kubenswrapper[4960]: E1201 19:41:35.324845 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:35 crc kubenswrapper[4960]: E1201 19:41:35.324925 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:35 crc kubenswrapper[4960]: E1201 19:41:35.324974 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:35 crc kubenswrapper[4960]: E1201 19:41:35.444384 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:41:37 crc kubenswrapper[4960]: I1201 19:41:37.323401 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:37 crc kubenswrapper[4960]: I1201 19:41:37.323490 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:37 crc kubenswrapper[4960]: I1201 19:41:37.323592 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:37 crc kubenswrapper[4960]: E1201 19:41:37.323617 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:37 crc kubenswrapper[4960]: I1201 19:41:37.323427 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:37 crc kubenswrapper[4960]: E1201 19:41:37.323803 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:37 crc kubenswrapper[4960]: E1201 19:41:37.323968 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:37 crc kubenswrapper[4960]: E1201 19:41:37.324101 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:38 crc kubenswrapper[4960]: I1201 19:41:38.324900 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.141157 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/3.log" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.145452 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerStarted","Data":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.145987 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.185049 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podStartSLOduration=111.185030937 podStartE2EDuration="1m51.185030937s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:39.184587494 +0000 UTC m=+134.472079163" watchObservedRunningTime="2025-12-01 19:41:39.185030937 +0000 UTC m=+134.472522606" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.323705 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.323789 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.323745 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.323721 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:39 crc kubenswrapper[4960]: E1201 19:41:39.323923 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:39 crc kubenswrapper[4960]: E1201 19:41:39.324092 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:39 crc kubenswrapper[4960]: E1201 19:41:39.324217 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:39 crc kubenswrapper[4960]: E1201 19:41:39.324287 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:39 crc kubenswrapper[4960]: I1201 19:41:39.358567 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jsbh7"] Dec 01 19:41:40 crc kubenswrapper[4960]: I1201 19:41:40.150230 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:40 crc kubenswrapper[4960]: E1201 19:41:40.150418 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:40 crc kubenswrapper[4960]: E1201 19:41:40.445913 4960 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:41:41 crc kubenswrapper[4960]: I1201 19:41:41.323604 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:41 crc kubenswrapper[4960]: I1201 19:41:41.323710 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:41 crc kubenswrapper[4960]: I1201 19:41:41.323734 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:41 crc kubenswrapper[4960]: I1201 19:41:41.323648 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:41 crc kubenswrapper[4960]: E1201 19:41:41.323880 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:41 crc kubenswrapper[4960]: E1201 19:41:41.324236 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:41 crc kubenswrapper[4960]: E1201 19:41:41.324329 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:41 crc kubenswrapper[4960]: E1201 19:41:41.324392 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:43 crc kubenswrapper[4960]: I1201 19:41:43.323871 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:43 crc kubenswrapper[4960]: E1201 19:41:43.323979 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:43 crc kubenswrapper[4960]: I1201 19:41:43.324058 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:43 crc kubenswrapper[4960]: I1201 19:41:43.324081 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:43 crc kubenswrapper[4960]: E1201 19:41:43.324261 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:43 crc kubenswrapper[4960]: I1201 19:41:43.324331 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:43 crc kubenswrapper[4960]: E1201 19:41:43.324538 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:43 crc kubenswrapper[4960]: E1201 19:41:43.324660 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:45 crc kubenswrapper[4960]: I1201 19:41:45.324014 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:45 crc kubenswrapper[4960]: I1201 19:41:45.324161 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:45 crc kubenswrapper[4960]: I1201 19:41:45.324232 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:45 crc kubenswrapper[4960]: I1201 19:41:45.324340 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:45 crc kubenswrapper[4960]: E1201 19:41:45.326441 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jsbh7" podUID="30d9bdd8-31e7-460b-aefb-421c6d306fbd" Dec 01 19:41:45 crc kubenswrapper[4960]: E1201 19:41:45.326572 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 01 19:41:45 crc kubenswrapper[4960]: E1201 19:41:45.326691 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 01 19:41:45 crc kubenswrapper[4960]: E1201 19:41:45.326844 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.323206 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.323297 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.323306 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.323229 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327202 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327346 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327347 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327496 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327538 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 19:41:47 crc kubenswrapper[4960]: I1201 19:41:47.327580 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.528671 4960 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.577757 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.578383 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.598434 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.598561 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.601507 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.602539 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.603600 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.603762 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.603950 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.612558 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.613287 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.614394 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.615295 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.615529 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.616094 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2w5q"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.617028 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.617098 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.617665 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.618265 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.618918 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.619596 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-g7thq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.619695 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.619936 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.619971 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.620296 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.620939 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mhbt7"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.622286 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.625594 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.625622 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.625866 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.625917 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.625863 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.626002 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.627206 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.627994 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.635806 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.636416 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.636710 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.636933 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.637301 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.637469 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638145 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638347 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638520 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638612 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-r5nl7"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638703 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638849 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638876 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.638994 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.639181 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.639446 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.639626 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.639845 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.639982 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.640009 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.640159 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.640301 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.647238 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.647625 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.647914 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.648420 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.648471 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.647641 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.653587 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5g7zf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.654237 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.654485 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v8sgq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.655899 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.656208 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.656233 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.656340 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.656529 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.656887 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.657385 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.657842 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.664264 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.664566 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.666007 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.666317 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.666439 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.666635 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.666918 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.667984 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.668466 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.669322 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.669445 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.670183 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.670439 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.670769 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.685321 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.685369 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.685482 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.685671 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.686339 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.686727 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.686987 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.688237 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.684309 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.689412 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.693426 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.696386 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.696632 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.696754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.696836 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.710498 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.711888 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.712866 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.713432 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.713701 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715107 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715308 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715404 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715453 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715498 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715642 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715765 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715775 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715931 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.715985 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.716017 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.716061 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.716169 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.716265 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.716349 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.717929 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718027 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718159 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-dpfnk"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718388 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718511 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718527 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718399 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718609 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718749 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718884 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.718990 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719042 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719068 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f66a36-f70e-4569-b679-9e676a2cd311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719101 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ce36e69-5e33-47d3-b526-41844fb4ff8e-metrics-tls\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719146 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-serving-cert\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719169 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1da67a53-d32d-4d47-aaf0-c9f43b48b863-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719383 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f66a36-f70e-4569-b679-9e676a2cd311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pc4f\" (UniqueName: \"kubernetes.io/projected/9d8d9a60-778d-4318-b05d-7958a905c66d-kube-api-access-6pc4f\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719534 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1da67a53-d32d-4d47-aaf0-c9f43b48b863-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719561 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de273fd5-2f0f-4532-b6ca-67cd9893ab26-serving-cert\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719584 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719601 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r2cq\" (UniqueName: \"kubernetes.io/projected/de273fd5-2f0f-4532-b6ca-67cd9893ab26-kube-api-access-8r2cq\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhm57\" (UniqueName: \"kubernetes.io/projected/19f66a36-f70e-4569-b679-9e676a2cd311-kube-api-access-bhm57\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719669 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2xls\" (UniqueName: \"kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719690 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-trusted-ca\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719714 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47hb5\" (UniqueName: \"kubernetes.io/projected/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-kube-api-access-47hb5\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719745 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-config\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719768 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d8d9a60-778d-4318-b05d-7958a905c66d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719798 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t85x\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-kube-api-access-8t85x\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719816 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgtzz\" (UniqueName: \"kubernetes.io/projected/7ce36e69-5e33-47d3-b526-41844fb4ff8e-kube-api-access-vgtzz\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719833 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/de273fd5-2f0f-4532-b6ca-67cd9893ab26-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.719867 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.721620 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.723990 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.724208 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.724327 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.724503 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.726035 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.726758 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.728332 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.729253 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.729754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.730506 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.734829 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.738858 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.741718 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.744060 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.744204 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.744346 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c48mv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.745914 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.746268 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.746315 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.747043 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.747224 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nz287"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.747901 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.748637 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.750900 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.752520 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.754111 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.754587 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.756464 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.757829 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5tw8"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.761262 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.762538 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.765508 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.766373 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.766346 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-w95l4"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.766561 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.768601 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.768760 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.769473 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.770266 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.770803 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.772183 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.773859 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9lvnl"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.774147 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.774915 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.775187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.775886 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.776689 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.778026 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.778554 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.779598 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.781034 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.781743 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.783249 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.785055 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-g7thq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.786336 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.787629 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.793007 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5g7zf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.796415 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v8sgq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.803401 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.805689 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mhbt7"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.808201 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.810174 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8cvhx"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.810992 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.812006 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2w5q"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.813186 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-r5nl7"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.814136 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.815504 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.816134 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.817570 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.817998 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.821438 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.823188 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.823918 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.823948 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.823971 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.823999 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pc4f\" (UniqueName: \"kubernetes.io/projected/9d8d9a60-778d-4318-b05d-7958a905c66d-kube-api-access-6pc4f\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824039 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c1d373-0f1b-4f75-bf87-eb6927a64389-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824074 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c1d373-0f1b-4f75-bf87-eb6927a64389-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1da67a53-d32d-4d47-aaf0-c9f43b48b863-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824159 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824177 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824194 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de273fd5-2f0f-4532-b6ca-67cd9893ab26-serving-cert\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824231 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r2cq\" (UniqueName: \"kubernetes.io/projected/de273fd5-2f0f-4532-b6ca-67cd9893ab26-kube-api-access-8r2cq\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824248 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824264 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-stats-auth\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824280 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824300 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhm57\" (UniqueName: \"kubernetes.io/projected/19f66a36-f70e-4569-b679-9e676a2cd311-kube-api-access-bhm57\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824321 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32c1d373-0f1b-4f75-bf87-eb6927a64389-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824337 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7w9b\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-kube-api-access-v7w9b\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-auth-proxy-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-default-certificate\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824387 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7gp6\" (UniqueName: \"kubernetes.io/projected/f084aac8-8a48-42db-b4bb-7773174cbd78-kube-api-access-j7gp6\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824403 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjr44\" (UniqueName: \"kubernetes.io/projected/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-kube-api-access-xjr44\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824420 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2xls\" (UniqueName: \"kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824439 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-trusted-ca\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-metrics-certs\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824472 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-config\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824526 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47hb5\" (UniqueName: \"kubernetes.io/projected/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-kube-api-access-47hb5\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824544 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f084aac8-8a48-42db-b4bb-7773174cbd78-service-ca-bundle\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824562 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824601 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t85x\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-kube-api-access-8t85x\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824640 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d8d9a60-778d-4318-b05d-7958a905c66d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824667 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824687 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgtzz\" (UniqueName: \"kubernetes.io/projected/7ce36e69-5e33-47d3-b526-41844fb4ff8e-kube-api-access-vgtzz\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824704 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb97s\" (UniqueName: \"kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824723 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/de273fd5-2f0f-4532-b6ca-67cd9893ab26-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824740 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824758 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824791 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824837 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824856 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824874 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f66a36-f70e-4569-b679-9e676a2cd311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8j9s\" (UniqueName: \"kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824911 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824929 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ce36e69-5e33-47d3-b526-41844fb4ff8e-metrics-tls\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824945 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-serving-cert\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824965 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1da67a53-d32d-4d47-aaf0-c9f43b48b863-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.824993 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.825010 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.825031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-machine-approver-tls\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.825048 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f66a36-f70e-4569-b679-9e676a2cd311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.825066 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.826245 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.828204 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/de273fd5-2f0f-4532-b6ca-67cd9893ab26-available-featuregates\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.828649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f66a36-f70e-4569-b679-9e676a2cd311-config\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.828776 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-trusted-ca\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.829965 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-config\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.830239 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.830430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.830785 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1da67a53-d32d-4d47-aaf0-c9f43b48b863-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.832873 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.834096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.834159 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.834433 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de273fd5-2f0f-4532-b6ca-67cd9893ab26-serving-cert\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.834841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19f66a36-f70e-4569-b679-9e676a2cd311-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.837154 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-serving-cert\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.837636 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7ce36e69-5e33-47d3-b526-41844fb4ff8e-metrics-tls\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.839017 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9d8d9a60-778d-4318-b05d-7958a905c66d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.839058 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9lvnl"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.847453 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.847489 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.847502 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.841563 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.852172 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.857368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1da67a53-d32d-4d47-aaf0-c9f43b48b863-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.859071 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c48mv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.860373 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.860862 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jgr5w"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.861471 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.861920 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.863917 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.864908 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nz287"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.866680 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.867836 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5tw8"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.869152 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.871741 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w95l4"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.874291 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jgr5w"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.875587 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wv5p9"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.876446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.877144 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7r8b"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.879562 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wv5p9"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.879609 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7r8b"] Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.879695 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.880423 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.900791 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.921244 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926073 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926139 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-machine-approver-tls\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926197 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926217 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926262 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c1d373-0f1b-4f75-bf87-eb6927a64389-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926289 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926321 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926349 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c1d373-0f1b-4f75-bf87-eb6927a64389-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926485 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926508 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-stats-auth\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926529 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926556 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32c1d373-0f1b-4f75-bf87-eb6927a64389-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7w9b\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-kube-api-access-v7w9b\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-auth-proxy-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926623 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-default-certificate\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926638 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7gp6\" (UniqueName: \"kubernetes.io/projected/f084aac8-8a48-42db-b4bb-7773174cbd78-kube-api-access-j7gp6\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926655 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjr44\" (UniqueName: \"kubernetes.io/projected/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-kube-api-access-xjr44\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926697 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926720 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-metrics-certs\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926750 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f084aac8-8a48-42db-b4bb-7773174cbd78-service-ca-bundle\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926770 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926789 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926848 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb97s\" (UniqueName: \"kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926890 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926934 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926953 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.926969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.927019 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8j9s\" (UniqueName: \"kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.927038 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.927520 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.928180 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.928326 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.928455 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.928656 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.929457 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.929497 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-auth-proxy-config\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.929863 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.930291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.930579 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-machine-approver-tls\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.930854 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.929097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c1d373-0f1b-4f75-bf87-eb6927a64389-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932032 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932231 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932341 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932631 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.932969 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c1d373-0f1b-4f75-bf87-eb6927a64389-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.933092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.933543 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.933981 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.934298 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.940572 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.948996 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.959650 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 19:41:51 crc kubenswrapper[4960]: I1201 19:41:51.980907 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.000358 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.026805 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.040094 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.041400 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.061219 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.082515 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.102221 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.123205 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.142190 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.161101 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.181680 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.202171 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.220903 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.241901 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.254995 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-metrics-tls\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.262023 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.292057 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.300814 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-trusted-ca\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.301707 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.321893 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.342378 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.356016 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-default-certificate\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.361629 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.375627 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-stats-auth\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.380929 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.394471 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f084aac8-8a48-42db-b4bb-7773174cbd78-metrics-certs\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.401200 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.411595 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f084aac8-8a48-42db-b4bb-7773174cbd78-service-ca-bundle\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.420875 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.460932 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.482292 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.501186 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.522013 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.536782 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.536991 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:52 crc kubenswrapper[4960]: E1201 19:41:52.537047 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:43:54.537000806 +0000 UTC m=+269.824492505 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.537204 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.537316 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.537357 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.539271 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.540608 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.543382 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.543634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.544654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.561426 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.581726 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.601561 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.621816 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.640959 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.662737 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.681574 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.702186 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.722700 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.742841 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.759390 4960 request.go:700] Waited for 1.005875506s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dmarketplace-trusted-ca&limit=500&resourceVersion=0 Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.765859 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.775615 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.779061 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.780715 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.789034 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.802553 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.822132 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.853725 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.863162 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.882718 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.903342 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.922525 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.941379 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.961220 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 19:41:52 crc kubenswrapper[4960]: I1201 19:41:52.981105 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.002505 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.021304 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 19:41:53 crc kubenswrapper[4960]: W1201 19:41:53.039389 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-063a3cf7e91509fcf953e24c9954383cafc1f4bca0feab805c8fc1242fe75a23 WatchSource:0}: Error finding container 063a3cf7e91509fcf953e24c9954383cafc1f4bca0feab805c8fc1242fe75a23: Status 404 returned error can't find the container with id 063a3cf7e91509fcf953e24c9954383cafc1f4bca0feab805c8fc1242fe75a23 Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.042463 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.061482 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.083086 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.101021 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.121829 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.142089 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.161938 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.185626 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.201663 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.214023 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"afe5daf72430ce0e3c2f7d40d83445ca5e1d062612fd6fedbf52529aab8e8b42"} Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.215923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"063a3cf7e91509fcf953e24c9954383cafc1f4bca0feab805c8fc1242fe75a23"} Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.217615 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ed929e3d2ae32d1530d47824c311210ae376a4eaeca734636f86ca43a69828dd"} Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.221502 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.241440 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.261464 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.282758 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.303535 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.323027 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.341737 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.361930 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.401927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.421868 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.441822 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.481920 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pc4f\" (UniqueName: \"kubernetes.io/projected/9d8d9a60-778d-4318-b05d-7958a905c66d-kube-api-access-6pc4f\") pod \"cluster-samples-operator-665b6dd947-p7vzg\" (UID: \"9d8d9a60-778d-4318-b05d-7958a905c66d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.488639 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.510964 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t85x\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-kube-api-access-8t85x\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.525697 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2xls\" (UniqueName: \"kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls\") pod \"route-controller-manager-6576b87f9c-dx2ds\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.547470 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r2cq\" (UniqueName: \"kubernetes.io/projected/de273fd5-2f0f-4532-b6ca-67cd9893ab26-kube-api-access-8r2cq\") pod \"openshift-config-operator-7777fb866f-6hdx9\" (UID: \"de273fd5-2f0f-4532-b6ca-67cd9893ab26\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.567895 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgtzz\" (UniqueName: \"kubernetes.io/projected/7ce36e69-5e33-47d3-b526-41844fb4ff8e-kube-api-access-vgtzz\") pod \"dns-operator-744455d44c-q2w5q\" (UID: \"7ce36e69-5e33-47d3-b526-41844fb4ff8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.584295 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1da67a53-d32d-4d47-aaf0-c9f43b48b863-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rcgdg\" (UID: \"1da67a53-d32d-4d47-aaf0-c9f43b48b863\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.613380 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhm57\" (UniqueName: \"kubernetes.io/projected/19f66a36-f70e-4569-b679-9e676a2cd311-kube-api-access-bhm57\") pod \"openshift-apiserver-operator-796bbdcf4f-k8b5p\" (UID: \"19f66a36-f70e-4569-b679-9e676a2cd311\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.625275 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.632156 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47hb5\" (UniqueName: \"kubernetes.io/projected/a1cbbec2-92a5-4a94-9932-7ff4f36b6860-kube-api-access-47hb5\") pod \"console-operator-58897d9998-g7thq\" (UID: \"a1cbbec2-92a5-4a94-9932-7ff4f36b6860\") " pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.642248 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.661392 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.681646 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.701470 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.724563 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.739701 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.741974 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.742345 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg"] Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.751214 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.761593 4960 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.779784 4960 request.go:700] Waited for 1.899829132s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.782167 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.801453 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.832277 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.842916 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.849522 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-bound-sa-token\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.849804 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.862943 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.870291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb97s\" (UniqueName: \"kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s\") pod \"console-f9d7485db-j9vkv\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.883137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/32c1d373-0f1b-4f75-bf87-eb6927a64389-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cj28l\" (UID: \"32c1d373-0f1b-4f75-bf87-eb6927a64389\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.902554 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7w9b\" (UniqueName: \"kubernetes.io/projected/9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4-kube-api-access-v7w9b\") pod \"ingress-operator-5b745b69d9-69vvk\" (UID: \"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.922484 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjr44\" (UniqueName: \"kubernetes.io/projected/cf21cb96-ba0f-4872-80bc-3f09db8f8ca8-kube-api-access-xjr44\") pod \"machine-approver-56656f9798-5vwj5\" (UID: \"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.940932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7gp6\" (UniqueName: \"kubernetes.io/projected/f084aac8-8a48-42db-b4bb-7773174cbd78-kube-api-access-j7gp6\") pod \"router-default-5444994796-dpfnk\" (UID: \"f084aac8-8a48-42db-b4bb-7773174cbd78\") " pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.958860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8j9s\" (UniqueName: \"kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s\") pod \"oauth-openshift-558db77b4-v8sgq\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969017 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969241 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-encryption-config\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969303 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-client\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/046442a5-907a-4147-8598-f5f836a2e082-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969346 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969382 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969461 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-serving-cert\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969493 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q4vg\" (UniqueName: \"kubernetes.io/projected/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-kube-api-access-9q4vg\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969516 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-config\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-image-import-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969616 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66243bfe-890f-4184-b937-b26e694adf33-serving-cert\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969656 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-node-pullsecrets\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-serving-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969827 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969844 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxkx\" (UniqueName: \"kubernetes.io/projected/314ac1b3-c601-4dda-8374-f76638dbc5cd-kube-api-access-hcxkx\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969912 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969928 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969950 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.969993 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970008 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-serving-cert\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970025 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970043 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-config\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970061 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqbkp\" (UniqueName: \"kubernetes.io/projected/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-kube-api-access-cqbkp\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970095 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970124 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-audit-dir\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970161 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970180 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-audit\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970208 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75z2r\" (UniqueName: \"kubernetes.io/projected/19bef78c-7725-4036-a57a-7e7335dbaed2-kube-api-access-75z2r\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970248 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: E1201 19:41:53.970751 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.47072952 +0000 UTC m=+149.758221189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970842 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046442a5-907a-4147-8598-f5f836a2e082-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970895 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb9gr\" (UniqueName: \"kubernetes.io/projected/046442a5-907a-4147-8598-f5f836a2e082-kube-api-access-fb9gr\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.970916 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-policies\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971135 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-service-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971165 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971303 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971328 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-images\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971373 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbf4h\" (UniqueName: \"kubernetes.io/projected/1b41ded7-08c7-4b04-bd13-a286d8a1bca4-kube-api-access-jbf4h\") pod \"migrator-59844c95c7-mhrdh\" (UID: \"1b41ded7-08c7-4b04-bd13-a286d8a1bca4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971487 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971581 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-client\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971628 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqxgn\" (UniqueName: \"kubernetes.io/projected/21ee4d43-3a2b-48a0-a5dd-332abfab285e-kube-api-access-lqxgn\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn4j7\" (UniqueName: \"kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971687 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chjc9\" (UniqueName: \"kubernetes.io/projected/66243bfe-890f-4184-b937-b26e694adf33-kube-api-access-chjc9\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971712 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-encryption-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971730 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-dir\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.971978 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh2kr\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.972006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-srv-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.978645 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:53 crc kubenswrapper[4960]: I1201 19:41:53.983592 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.007644 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.018127 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.025098 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.025788 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.073868 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.074556 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-audit\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.075436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-audit\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.075528 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75z2r\" (UniqueName: \"kubernetes.io/projected/19bef78c-7725-4036-a57a-7e7335dbaed2-kube-api-access-75z2r\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.075580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.075865 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.57584833 +0000 UTC m=+149.863339999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.075896 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046442a5-907a-4147-8598-f5f836a2e082-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.075963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfmf9\" (UniqueName: \"kubernetes.io/projected/b8af7379-3edf-45ef-9564-fe4f6cbf8561-kube-api-access-gfmf9\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.076598 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046442a5-907a-4147-8598-f5f836a2e082-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.076881 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.076948 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-images\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.076967 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7txkf\" (UniqueName: \"kubernetes.io/projected/dd5a0cdf-1451-4787-b68b-98b72c5173d7-kube-api-access-7txkf\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078416 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-images\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078544 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbf4h\" (UniqueName: \"kubernetes.io/projected/1b41ded7-08c7-4b04-bd13-a286d8a1bca4-kube-api-access-jbf4h\") pod \"migrator-59844c95c7-mhrdh\" (UID: \"1b41ded7-08c7-4b04-bd13-a286d8a1bca4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078584 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078657 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-client\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078771 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqxgn\" (UniqueName: \"kubernetes.io/projected/21ee4d43-3a2b-48a0-a5dd-332abfab285e-kube-api-access-lqxgn\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.078901 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-srv-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.080253 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.099292 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-client\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.100780 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.100853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.100890 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.100913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.101074 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-serving-cert\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.101135 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66243bfe-890f-4184-b937-b26e694adf33-serving-cert\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.102575 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.103191 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.104716 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105251 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-srv-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105390 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-serving-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105546 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx92x\" (UniqueName: \"kubernetes.io/projected/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-kube-api-access-nx92x\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105602 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105641 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.105800 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.106592 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-serving-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.106643 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.106944 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-serving-cert\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.107128 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.107941 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.108521 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.107022 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111277 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd5a0cdf-1451-4787-b68b-98b72c5173d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-config\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111413 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-config\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111437 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111480 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111507 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-audit-dir\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111664 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.111756 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-cert\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.112595 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.112654 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb9gr\" (UniqueName: \"kubernetes.io/projected/046442a5-907a-4147-8598-f5f836a2e082-kube-api-access-fb9gr\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.112693 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.112706 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.113757 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-service-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.113905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114022 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-policies\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114319 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114558 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn4j7\" (UniqueName: \"kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114607 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-encryption-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114636 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-dir\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.114690 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chjc9\" (UniqueName: \"kubernetes.io/projected/66243bfe-890f-4184-b937-b26e694adf33-kube-api-access-chjc9\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.115645 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.116493 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-audit-dir\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.116553 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.116805 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-policies\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.116889 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh2kr\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.117385 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.117471 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/21ee4d43-3a2b-48a0-a5dd-332abfab285e-audit-dir\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.117856 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.617828235 +0000 UTC m=+149.905319904 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.118509 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-encryption-config\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.118656 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.118758 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-client\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.118890 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/046442a5-907a-4147-8598-f5f836a2e082-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.121327 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.122027 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-service-ca-bundle\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.122769 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66243bfe-890f-4184-b937-b26e694adf33-config\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.125331 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q4vg\" (UniqueName: \"kubernetes.io/projected/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-kube-api-access-9q4vg\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.125623 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-config\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.125727 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.125820 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-image-import-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.126696 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-node-pullsecrets\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.126783 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8af7379-3edf-45ef-9564-fe4f6cbf8561-proxy-tls\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.126862 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-images\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.126915 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/19bef78c-7725-4036-a57a-7e7335dbaed2-image-import-ca\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.126941 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.127193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/19bef78c-7725-4036-a57a-7e7335dbaed2-node-pullsecrets\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.127232 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-config\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.127018 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxkx\" (UniqueName: \"kubernetes.io/projected/314ac1b3-c601-4dda-8374-f76638dbc5cd-kube-api-access-hcxkx\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.128079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.128148 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqbkp\" (UniqueName: \"kubernetes.io/projected/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-kube-api-access-cqbkp\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.128932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/046442a5-907a-4147-8598-f5f836a2e082-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.137025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.139306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqxgn\" (UniqueName: \"kubernetes.io/projected/21ee4d43-3a2b-48a0-a5dd-332abfab285e-kube-api-access-lqxgn\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.145258 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-serving-cert\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.147889 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.152728 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75z2r\" (UniqueName: \"kubernetes.io/projected/19bef78c-7725-4036-a57a-7e7335dbaed2-kube-api-access-75z2r\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.161348 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.167871 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/21ee4d43-3a2b-48a0-a5dd-332abfab285e-encryption-config\") pod \"apiserver-7bbb656c7d-m5ktp\" (UID: \"21ee4d43-3a2b-48a0-a5dd-332abfab285e\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.168678 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-serving-cert\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.168932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-encryption-config\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.169390 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/314ac1b3-c601-4dda-8374-f76638dbc5cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.169652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/19bef78c-7725-4036-a57a-7e7335dbaed2-etcd-client\") pod \"apiserver-76f77b778f-5g7zf\" (UID: \"19bef78c-7725-4036-a57a-7e7335dbaed2\") " pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.169831 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66243bfe-890f-4184-b937-b26e694adf33-serving-cert\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.170429 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.179832 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbf4h\" (UniqueName: \"kubernetes.io/projected/1b41ded7-08c7-4b04-bd13-a286d8a1bca4-kube-api-access-jbf4h\") pod \"migrator-59844c95c7-mhrdh\" (UID: \"1b41ded7-08c7-4b04-bd13-a286d8a1bca4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.179872 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.189711 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.196435 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.200377 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.212392 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q2w5q"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.225284 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.230801 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" event={"ID":"156be4ee-5caf-4423-82aa-9607f9abc19d","Type":"ContainerStarted","Data":"d78576447b25c7ab2ac76890745d2317eeee7f647095dd0e3803252c40d02479"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.233607 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1408afc68a7cf97cc7d94899ae9d413456f0895bd8ad6c86f1ae3b9fad3a336e"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.235996 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb9gr\" (UniqueName: \"kubernetes.io/projected/046442a5-907a-4147-8598-f5f836a2e082-kube-api-access-fb9gr\") pod \"kube-storage-version-migrator-operator-b67b599dd-ms5tj\" (UID: \"046442a5-907a-4147-8598-f5f836a2e082\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236500 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f36c6ade923db1a812bf9895e705d78890bcf5ae969b41adcb8a931f5c697245"} Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.236797 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.736775836 +0000 UTC m=+150.024267515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236836 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2jtq\" (UniqueName: \"kubernetes.io/projected/ad5cd286-cc5e-4c29-a98c-992994d3ada2-kube-api-access-d2jtq\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236863 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2pqm\" (UniqueName: \"kubernetes.io/projected/ffd93427-d545-4290-830e-329e8b42dce6-kube-api-access-m2pqm\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3386a2b-dc67-4f1d-ba89-04e29243819b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236924 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j86c\" (UniqueName: \"kubernetes.io/projected/7a8f833e-a2ef-4058-9ba8-99408a789116-kube-api-access-9j86c\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.236968 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/116b4f38-75e4-4b24-8e46-41e9d78a8c81-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237001 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-cert\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237036 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237063 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237089 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237127 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7a8f833e-a2ef-4058-9ba8-99408a789116-tmpfs\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237161 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gxs5\" (UniqueName: \"kubernetes.io/projected/c30ab298-c528-45be-99af-d8d1135eef2f-kube-api-access-8gxs5\") pod \"downloads-7954f5f757-w95l4\" (UID: \"c30ab298-c528-45be-99af-d8d1135eef2f\") " pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237218 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-socket-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237250 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-certs\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237280 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-csi-data-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237306 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237353 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-plugins-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237379 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7sd5\" (UniqueName: \"kubernetes.io/projected/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-kube-api-access-t7sd5\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237405 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3386a2b-dc67-4f1d-ba89-04e29243819b-config\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237431 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4qfn\" (UniqueName: \"kubernetes.io/projected/6a734d24-91a7-4d3c-a719-403b9aa87c9a-kube-api-access-l4qfn\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-serving-cert\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237475 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-registration-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-cabundle\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237519 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-mountpoint-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/116b4f38-75e4-4b24-8e46-41e9d78a8c81-proxy-tls\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8af7379-3edf-45ef-9564-fe4f6cbf8561-proxy-tls\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-images\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237622 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-node-bootstrap-token\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237652 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6qgc\" (UniqueName: \"kubernetes.io/projected/84c9d692-2ab6-42de-b2e9-320ae0561090-kube-api-access-j6qgc\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237681 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9kvg\" (UniqueName: \"kubernetes.io/projected/0db6e6bb-0673-416b-a48d-4cfa433893bb-kube-api-access-s9kvg\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237702 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-config\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237726 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237752 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjlhb\" (UniqueName: \"kubernetes.io/projected/9b8a78a1-d86c-498b-8bd2-52ff3687455e-kube-api-access-bjlhb\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237793 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237817 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-webhook-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237840 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a734d24-91a7-4d3c-a719-403b9aa87c9a-config\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfmf9\" (UniqueName: \"kubernetes.io/projected/b8af7379-3edf-45ef-9564-fe4f6cbf8561-kube-api-access-gfmf9\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237918 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdxpg\" (UniqueName: \"kubernetes.io/projected/952f446b-a8d2-4887-bc67-cddfb9e20791-kube-api-access-qdxpg\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.237967 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7txkf\" (UniqueName: \"kubernetes.io/projected/dd5a0cdf-1451-4787-b68b-98b72c5173d7-kube-api-access-7txkf\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.238574 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad5cd286-cc5e-4c29-a98c-992994d3ada2-config-volume\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.239460 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-key\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.239564 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsqrt\" (UniqueName: \"kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.239593 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffd93427-d545-4290-830e-329e8b42dce6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.239620 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-srv-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.239971 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a734d24-91a7-4d3c-a719-403b9aa87c9a-serving-cert\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.240291 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.240319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-client\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.240342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/952f446b-a8d2-4887-bc67-cddfb9e20791-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.241410 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.741387586 +0000 UTC m=+150.028879255 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.242064 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b8af7379-3edf-45ef-9564-fe4f6cbf8561-proxy-tls\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.242656 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-images\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.243671 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.243708 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-service-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244292 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b8af7379-3edf-45ef-9564-fe4f6cbf8561-auth-proxy-config\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244352 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad5cd286-cc5e-4c29-a98c-992994d3ada2-metrics-tls\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244377 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-apiservice-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244425 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx92x\" (UniqueName: \"kubernetes.io/projected/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-kube-api-access-nx92x\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244467 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4xwh\" (UniqueName: \"kubernetes.io/projected/ff6d1596-2e90-47bb-a6ff-4335b49328f6-kube-api-access-n4xwh\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.244488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw2gt\" (UniqueName: \"kubernetes.io/projected/116b4f38-75e4-4b24-8e46-41e9d78a8c81-kube-api-access-tw2gt\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247166 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-cert\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247508 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3386a2b-dc67-4f1d-ba89-04e29243819b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247631 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd5a0cdf-1451-4787-b68b-98b72c5173d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247684 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247771 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.247931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-config\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.248087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" event={"ID":"19f66a36-f70e-4569-b679-9e676a2cd311","Type":"ContainerStarted","Data":"2f9860b9fac927d4c42a840e497e88403084024742b848308e80de77d2b5789e"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.248144 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mndrf\" (UniqueName: \"kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.249621 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-config\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.251376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn4j7\" (UniqueName: \"kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7\") pod \"controller-manager-879f6c89f-q28sf\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.251440 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.253932 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" event={"ID":"9d8d9a60-778d-4318-b05d-7958a905c66d","Type":"ContainerStarted","Data":"600e9ef1fbb3d0180794334087093fc9e811b27a8ad2d4e3372765b67394efa1"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.254017 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" event={"ID":"9d8d9a60-778d-4318-b05d-7958a905c66d","Type":"ContainerStarted","Data":"68079060f530bdb9ae048d3b7a1a90af2518a6a75d14940dda80aa35741ed461"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.256524 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.258064 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dpfnk" event={"ID":"f084aac8-8a48-42db-b4bb-7773174cbd78","Type":"ContainerStarted","Data":"a1bcb90ef4d1d7e74f5bd887e414dd01ab380de1c55f3ec7ae886edd83e46260"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.262527 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ecd130bbed7a262187fecfbf61a370c97214a6a75e192c37db5eee15b98b1ca4"} Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.262722 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.267250 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd5a0cdf-1451-4787-b68b-98b72c5173d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.278534 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh2kr\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.280792 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chjc9\" (UniqueName: \"kubernetes.io/projected/66243bfe-890f-4184-b937-b26e694adf33-kube-api-access-chjc9\") pod \"authentication-operator-69f744f599-r5nl7\" (UID: \"66243bfe-890f-4184-b937-b26e694adf33\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.289992 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.296901 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.296920 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q4vg\" (UniqueName: \"kubernetes.io/projected/54cb7d4f-88bd-4a3b-a87b-3514e9babca6-kube-api-access-9q4vg\") pod \"openshift-controller-manager-operator-756b6f6bc6-8p4zf\" (UID: \"54cb7d4f-88bd-4a3b-a87b-3514e9babca6\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.311744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.334766 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxkx\" (UniqueName: \"kubernetes.io/projected/314ac1b3-c601-4dda-8374-f76638dbc5cd-kube-api-access-hcxkx\") pod \"catalog-operator-68c6474976-qdx9p\" (UID: \"314ac1b3-c601-4dda-8374-f76638dbc5cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.341919 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqbkp\" (UniqueName: \"kubernetes.io/projected/b18c64d6-cd4c-4d07-9591-4c9868b33b2c-kube-api-access-cqbkp\") pod \"machine-api-operator-5694c8668f-mhbt7\" (UID: \"b18c64d6-cd4c-4d07-9591-4c9868b33b2c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352280 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352689 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-socket-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352721 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-certs\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352790 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-csi-data-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352815 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3386a2b-dc67-4f1d-ba89-04e29243819b-config\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352854 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4qfn\" (UniqueName: \"kubernetes.io/projected/6a734d24-91a7-4d3c-a719-403b9aa87c9a-kube-api-access-l4qfn\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352881 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-plugins-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352907 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7sd5\" (UniqueName: \"kubernetes.io/projected/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-kube-api-access-t7sd5\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352951 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-serving-cert\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.352974 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-registration-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353016 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-cabundle\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-mountpoint-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353067 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/116b4f38-75e4-4b24-8e46-41e9d78a8c81-proxy-tls\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353109 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-node-bootstrap-token\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353157 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6qgc\" (UniqueName: \"kubernetes.io/projected/84c9d692-2ab6-42de-b2e9-320ae0561090-kube-api-access-j6qgc\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353181 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9kvg\" (UniqueName: \"kubernetes.io/projected/0db6e6bb-0673-416b-a48d-4cfa433893bb-kube-api-access-s9kvg\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-config\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353452 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjlhb\" (UniqueName: \"kubernetes.io/projected/9b8a78a1-d86c-498b-8bd2-52ff3687455e-kube-api-access-bjlhb\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353478 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353537 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-webhook-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353558 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a734d24-91a7-4d3c-a719-403b9aa87c9a-config\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353603 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdxpg\" (UniqueName: \"kubernetes.io/projected/952f446b-a8d2-4887-bc67-cddfb9e20791-kube-api-access-qdxpg\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.353627 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad5cd286-cc5e-4c29-a98c-992994d3ada2-config-volume\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356138 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-key\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsqrt\" (UniqueName: \"kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356201 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-srv-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffd93427-d545-4290-830e-329e8b42dce6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356882 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a734d24-91a7-4d3c-a719-403b9aa87c9a-serving-cert\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356915 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356956 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-client\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.356987 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/952f446b-a8d2-4887-bc67-cddfb9e20791-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-service-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357061 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad5cd286-cc5e-4c29-a98c-992994d3ada2-metrics-tls\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-apiservice-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357209 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4xwh\" (UniqueName: \"kubernetes.io/projected/ff6d1596-2e90-47bb-a6ff-4335b49328f6-kube-api-access-n4xwh\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357230 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw2gt\" (UniqueName: \"kubernetes.io/projected/116b4f38-75e4-4b24-8e46-41e9d78a8c81-kube-api-access-tw2gt\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357247 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3386a2b-dc67-4f1d-ba89-04e29243819b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357304 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357325 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mndrf\" (UniqueName: \"kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357364 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2jtq\" (UniqueName: \"kubernetes.io/projected/ad5cd286-cc5e-4c29-a98c-992994d3ada2-kube-api-access-d2jtq\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357381 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2pqm\" (UniqueName: \"kubernetes.io/projected/ffd93427-d545-4290-830e-329e8b42dce6-kube-api-access-m2pqm\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/116b4f38-75e4-4b24-8e46-41e9d78a8c81-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357418 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3386a2b-dc67-4f1d-ba89-04e29243819b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357451 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j86c\" (UniqueName: \"kubernetes.io/projected/7a8f833e-a2ef-4058-9ba8-99408a789116-kube-api-access-9j86c\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357478 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357495 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7a8f833e-a2ef-4058-9ba8-99408a789116-tmpfs\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357524 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.357549 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gxs5\" (UniqueName: \"kubernetes.io/projected/c30ab298-c528-45be-99af-d8d1135eef2f-kube-api-access-8gxs5\") pod \"downloads-7954f5f757-w95l4\" (UID: \"c30ab298-c528-45be-99af-d8d1135eef2f\") " pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.359659 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-config\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.359834 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.859820261 +0000 UTC m=+150.147311930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.361165 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-registration-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.361242 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-plugins-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.361718 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.361832 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-node-bootstrap-token\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.361913 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-mountpoint-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.362034 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-socket-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.363404 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-cabundle\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.363472 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-csi-data-dir\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.364806 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a734d24-91a7-4d3c-a719-403b9aa87c9a-config\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.366540 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/84c9d692-2ab6-42de-b2e9-320ae0561090-certs\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.368140 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3386a2b-dc67-4f1d-ba89-04e29243819b-config\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.374777 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-serving-cert\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.375303 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-service-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.379281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.380680 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad5cd286-cc5e-4c29-a98c-992994d3ada2-config-volume\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.380786 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7a8f833e-a2ef-4058-9ba8-99408a789116-tmpfs\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.381378 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-ca\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.384539 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-webhook-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.385107 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/116b4f38-75e4-4b24-8e46-41e9d78a8c81-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.387232 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ffd93427-d545-4290-830e-329e8b42dce6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.388747 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3386a2b-dc67-4f1d-ba89-04e29243819b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.388931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ff6d1596-2e90-47bb-a6ff-4335b49328f6-etcd-client\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.389044 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/116b4f38-75e4-4b24-8e46-41e9d78a8c81-proxy-tls\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.393938 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ad5cd286-cc5e-4c29-a98c-992994d3ada2-metrics-tls\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.394929 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7a8f833e-a2ef-4058-9ba8-99408a789116-apiservice-cert\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.395221 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.395244 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.395395 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/0db6e6bb-0673-416b-a48d-4cfa433893bb-signing-key\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.396073 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.396094 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/952f446b-a8d2-4887-bc67-cddfb9e20791-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.396406 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9b8a78a1-d86c-498b-8bd2-52ff3687455e-srv-cert\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.402477 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a734d24-91a7-4d3c-a719-403b9aa87c9a-serving-cert\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.406985 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfmf9\" (UniqueName: \"kubernetes.io/projected/b8af7379-3edf-45ef-9564-fe4f6cbf8561-kube-api-access-gfmf9\") pod \"machine-config-operator-74547568cd-sll8n\" (UID: \"b8af7379-3edf-45ef-9564-fe4f6cbf8561\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.415861 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7txkf\" (UniqueName: \"kubernetes.io/projected/dd5a0cdf-1451-4787-b68b-98b72c5173d7-kube-api-access-7txkf\") pod \"multus-admission-controller-857f4d67dd-z5tw8\" (UID: \"dd5a0cdf-1451-4787-b68b-98b72c5173d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.420362 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf0fccd8-8b4e-422e-abdc-ecf769334f1d-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-2xqzt\" (UID: \"bf0fccd8-8b4e-422e-abdc-ecf769334f1d\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.420782 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.421221 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.433589 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.460659 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.461015 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:54.961002973 +0000 UTC m=+150.248494642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.462982 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gxs5\" (UniqueName: \"kubernetes.io/projected/c30ab298-c528-45be-99af-d8d1135eef2f-kube-api-access-8gxs5\") pod \"downloads-7954f5f757-w95l4\" (UID: \"c30ab298-c528-45be-99af-d8d1135eef2f\") " pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.475091 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.476339 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx92x\" (UniqueName: \"kubernetes.io/projected/988d62ee-9dbc-44a4-b67e-6a2baeb2df7d-kube-api-access-nx92x\") pod \"ingress-canary-jgr5w\" (UID: \"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d\") " pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.483705 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6qgc\" (UniqueName: \"kubernetes.io/projected/84c9d692-2ab6-42de-b2e9-320ae0561090-kube-api-access-j6qgc\") pod \"machine-config-server-8cvhx\" (UID: \"84c9d692-2ab6-42de-b2e9-320ae0561090\") " pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.490206 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8cvhx" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.502466 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jgr5w" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.503446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9kvg\" (UniqueName: \"kubernetes.io/projected/0db6e6bb-0673-416b-a48d-4cfa433893bb-kube-api-access-s9kvg\") pod \"service-ca-9c57cc56f-9lvnl\" (UID: \"0db6e6bb-0673-416b-a48d-4cfa433893bb\") " pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.509567 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.514460 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.516820 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjlhb\" (UniqueName: \"kubernetes.io/projected/9b8a78a1-d86c-498b-8bd2-52ff3687455e-kube-api-access-bjlhb\") pod \"olm-operator-6b444d44fb-sgctq\" (UID: \"9b8a78a1-d86c-498b-8bd2-52ff3687455e\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.528649 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.537966 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.538511 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.540042 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4qfn\" (UniqueName: \"kubernetes.io/projected/6a734d24-91a7-4d3c-a719-403b9aa87c9a-kube-api-access-l4qfn\") pod \"service-ca-operator-777779d784-6xgg5\" (UID: \"6a734d24-91a7-4d3c-a719-403b9aa87c9a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.552917 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-g7thq"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.565683 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.566209 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.066184416 +0000 UTC m=+150.353676085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.568745 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7sd5\" (UniqueName: \"kubernetes.io/projected/4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b-kube-api-access-t7sd5\") pod \"csi-hostpathplugin-z7r8b\" (UID: \"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b\") " pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.587168 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdxpg\" (UniqueName: \"kubernetes.io/projected/952f446b-a8d2-4887-bc67-cddfb9e20791-kube-api-access-qdxpg\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qhvh\" (UID: \"952f446b-a8d2-4887-bc67-cddfb9e20791\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: W1201 19:41:54.591925 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee9aff31_fb6f_41a7_8cf1_3f6b549ec4a0.slice/crio-f7aa4d6f4a853d744095b449ea12619f9780956d202316f2abd8fac1e4a697df WatchSource:0}: Error finding container f7aa4d6f4a853d744095b449ea12619f9780956d202316f2abd8fac1e4a697df: Status 404 returned error can't find the container with id f7aa4d6f4a853d744095b449ea12619f9780956d202316f2abd8fac1e4a697df Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.606051 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.618397 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsqrt\" (UniqueName: \"kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt\") pod \"collect-profiles-29410290-c9bfb\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.626252 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2jtq\" (UniqueName: \"kubernetes.io/projected/ad5cd286-cc5e-4c29-a98c-992994d3ada2-kube-api-access-d2jtq\") pod \"dns-default-wv5p9\" (UID: \"ad5cd286-cc5e-4c29-a98c-992994d3ada2\") " pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: W1201 19:41:54.634298 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1cbbec2_92a5_4a94_9932_7ff4f36b6860.slice/crio-e1cfcdc2afdfb3d746219bdd40e727e93712c400cfa5e74bbde76f8ab4ff166d WatchSource:0}: Error finding container e1cfcdc2afdfb3d746219bdd40e727e93712c400cfa5e74bbde76f8ab4ff166d: Status 404 returned error can't find the container with id e1cfcdc2afdfb3d746219bdd40e727e93712c400cfa5e74bbde76f8ab4ff166d Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.636327 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5g7zf"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.656264 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.661457 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.665566 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mndrf\" (UniqueName: \"kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf\") pod \"marketplace-operator-79b997595-w6wcz\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.669148 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.669628 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.169614025 +0000 UTC m=+150.457105694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.673091 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v8sgq"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.686633 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j86c\" (UniqueName: \"kubernetes.io/projected/7a8f833e-a2ef-4058-9ba8-99408a789116-kube-api-access-9j86c\") pod \"packageserver-d55dfcdfc-qx5vs\" (UID: \"7a8f833e-a2ef-4058-9ba8-99408a789116\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.687628 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.698424 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.701386 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4xwh\" (UniqueName: \"kubernetes.io/projected/ff6d1596-2e90-47bb-a6ff-4335b49328f6-kube-api-access-n4xwh\") pod \"etcd-operator-b45778765-c48mv\" (UID: \"ff6d1596-2e90-47bb-a6ff-4335b49328f6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.701899 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.712107 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.715733 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:54 crc kubenswrapper[4960]: W1201 19:41:54.717326 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19bef78c_7725_4036_a57a_7e7335dbaed2.slice/crio-e2d61d8f6fd3419567863ff06aadfa87a4e517c12c2432225e1b4571870716dd WatchSource:0}: Error finding container e2d61d8f6fd3419567863ff06aadfa87a4e517c12c2432225e1b4571870716dd: Status 404 returned error can't find the container with id e2d61d8f6fd3419567863ff06aadfa87a4e517c12c2432225e1b4571870716dd Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.724619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2pqm\" (UniqueName: \"kubernetes.io/projected/ffd93427-d545-4290-830e-329e8b42dce6-kube-api-access-m2pqm\") pod \"package-server-manager-789f6589d5-j8xsm\" (UID: \"ffd93427-d545-4290-830e-329e8b42dce6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.744393 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.744996 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.754623 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.764599 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw2gt\" (UniqueName: \"kubernetes.io/projected/116b4f38-75e4-4b24-8e46-41e9d78a8c81-kube-api-access-tw2gt\") pod \"machine-config-controller-84d6567774-nz287\" (UID: \"116b4f38-75e4-4b24-8e46-41e9d78a8c81\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.770141 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3386a2b-dc67-4f1d-ba89-04e29243819b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-66whs\" (UID: \"a3386a2b-dc67-4f1d-ba89-04e29243819b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.772234 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.773591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.774355 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.274329745 +0000 UTC m=+150.561821424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.782525 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.804873 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh"] Dec 01 19:41:54 crc kubenswrapper[4960]: W1201 19:41:54.808016 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod046442a5_907a_4147_8598_f5f836a2e082.slice/crio-733c68b7314b4891ec995c921ba98acdc7abff025cce10597c2041b7ff7cd186 WatchSource:0}: Error finding container 733c68b7314b4891ec995c921ba98acdc7abff025cce10597c2041b7ff7cd186: Status 404 returned error can't find the container with id 733c68b7314b4891ec995c921ba98acdc7abff025cce10597c2041b7ff7cd186 Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.821781 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.850691 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.879182 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.880280 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.38026361 +0000 UTC m=+150.667755279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.898010 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n"] Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.966140 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.974710 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.980086 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.980697 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.480668238 +0000 UTC m=+150.768159907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.980886 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:54 crc kubenswrapper[4960]: E1201 19:41:54.981627 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.481613017 +0000 UTC m=+150.769104686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.982821 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" Dec 01 19:41:54 crc kubenswrapper[4960]: I1201 19:41:54.995090 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt"] Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.082640 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.082992 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.582972044 +0000 UTC m=+150.870463713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: W1201 19:41:55.154291 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf0fccd8_8b4e_422e_abdc_ecf769334f1d.slice/crio-fbe0a76c143ac0196d179136da5463641a5ae5da3ef3533087f5ed65f691577c WatchSource:0}: Error finding container fbe0a76c143ac0196d179136da5463641a5ae5da3ef3533087f5ed65f691577c: Status 404 returned error can't find the container with id fbe0a76c143ac0196d179136da5463641a5ae5da3ef3533087f5ed65f691577c Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.192667 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.193167 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.693145648 +0000 UTC m=+150.980637317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.269421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" event={"ID":"32c1d373-0f1b-4f75-bf87-eb6927a64389","Type":"ContainerStarted","Data":"641979c043f4183ae65cf61e556bd832fa31208b3eff4a1e5570c7b72765d5b3"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.270975 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-g7thq" event={"ID":"a1cbbec2-92a5-4a94-9932-7ff4f36b6860","Type":"ContainerStarted","Data":"e1cfcdc2afdfb3d746219bdd40e727e93712c400cfa5e74bbde76f8ab4ff166d"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.272469 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" event={"ID":"cc354d2f-b7ad-4878-8a1d-ffae0d774db3","Type":"ContainerStarted","Data":"97ac7a3e8707ea5f3c4971ac5e79f040c439749bd1b8f582e8525027b83c0ab4"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.273924 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" event={"ID":"7ce36e69-5e33-47d3-b526-41844fb4ff8e","Type":"ContainerStarted","Data":"592e17043f4c2b32730bd3551e3cccfbcef5982878b084d2cc71faff14984b60"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.273965 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" event={"ID":"7ce36e69-5e33-47d3-b526-41844fb4ff8e","Type":"ContainerStarted","Data":"07ca3209cae89e8347a74dc5e9afd65e060fc0a5deec4a99df23ec13d51b4145"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.274894 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" event={"ID":"21ee4d43-3a2b-48a0-a5dd-332abfab285e","Type":"ContainerStarted","Data":"0624ab3e33d02e70dffd50b2d8f9dcf1631948d621da5b929b2349847161132f"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.275822 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" event={"ID":"bf0fccd8-8b4e-422e-abdc-ecf769334f1d","Type":"ContainerStarted","Data":"fbe0a76c143ac0196d179136da5463641a5ae5da3ef3533087f5ed65f691577c"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.277182 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" event={"ID":"156be4ee-5caf-4423-82aa-9607f9abc19d","Type":"ContainerStarted","Data":"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.277446 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.278828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" event={"ID":"1da67a53-d32d-4d47-aaf0-c9f43b48b863","Type":"ContainerStarted","Data":"f5ca5dc82500f5d50620f110a84dcd650ba0e9a5ab68752235560c9d35447ea8"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.278886 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" event={"ID":"1da67a53-d32d-4d47-aaf0-c9f43b48b863","Type":"ContainerStarted","Data":"0c3477ea2af42df428aa2288908ba2f5b78d34a7fb3c5deebf8b9dfc0db9e884"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.281495 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8cvhx" event={"ID":"84c9d692-2ab6-42de-b2e9-320ae0561090","Type":"ContainerStarted","Data":"adfe460a94015037eb0fedc99a56444927dde1e2dcf85def69c531b2e5dc9ba9"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.284341 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" event={"ID":"19f66a36-f70e-4569-b679-9e676a2cd311","Type":"ContainerStarted","Data":"bef83169fed6cebcf9dc7225eabebeb4dac4b94aa1f4e761a343683e07f5a658"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.284495 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.285852 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" event={"ID":"046442a5-907a-4147-8598-f5f836a2e082","Type":"ContainerStarted","Data":"733c68b7314b4891ec995c921ba98acdc7abff025cce10597c2041b7ff7cd186"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.287816 4960 generic.go:334] "Generic (PLEG): container finished" podID="de273fd5-2f0f-4532-b6ca-67cd9893ab26" containerID="eeacd70b07b401f31d1700d2ccc0583e616ac70384f7ccef46c3cdf5499003ff" exitCode=0 Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.287921 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" event={"ID":"de273fd5-2f0f-4532-b6ca-67cd9893ab26","Type":"ContainerDied","Data":"eeacd70b07b401f31d1700d2ccc0583e616ac70384f7ccef46c3cdf5499003ff"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.287991 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" event={"ID":"de273fd5-2f0f-4532-b6ca-67cd9893ab26","Type":"ContainerStarted","Data":"f0060f327edb0897bf7c6bfe66dfea961aa0b85950da2b351be1617af714b42e"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.288929 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" event={"ID":"b8af7379-3edf-45ef-9564-fe4f6cbf8561","Type":"ContainerStarted","Data":"43050c245030fddaf1261ed19f0eb095ea68f6225b0761e25973c9fcf225f4b8"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.289945 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j9vkv" event={"ID":"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0","Type":"ContainerStarted","Data":"f7aa4d6f4a853d744095b449ea12619f9780956d202316f2abd8fac1e4a697df"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.291058 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dpfnk" event={"ID":"f084aac8-8a48-42db-b4bb-7773174cbd78","Type":"ContainerStarted","Data":"617b325b50f9bbcfe2859badc97d941ce25c7a53454a11a0c54f7c96da3a9fcd"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.292213 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" event={"ID":"1b41ded7-08c7-4b04-bd13-a286d8a1bca4","Type":"ContainerStarted","Data":"e5eb9e72bcc40ed8c18ce1670ef82ab4f6f7079591c2a07f02c8d220af4785c2"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.293230 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.293561 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" event={"ID":"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4","Type":"ContainerStarted","Data":"b7acb4e7257b5875e119f0ba113fbf8f4414be679abd48f73f35eeb907a75303"} Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.293615 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.793591478 +0000 UTC m=+151.081083147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.293843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.294302 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.794285429 +0000 UTC m=+151.081777098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.295138 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" event={"ID":"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8","Type":"ContainerStarted","Data":"3d4377bdb11afc328f8042cc95764fd229b80ba33a7ceb7dde014fc0d8b8d174"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.295180 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" event={"ID":"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8","Type":"ContainerStarted","Data":"db0709b2b910ef167584d9545e521a5359cfc0f69363cba961f4178d37821763"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.296271 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" event={"ID":"19bef78c-7725-4036-a57a-7e7335dbaed2","Type":"ContainerStarted","Data":"e2d61d8f6fd3419567863ff06aadfa87a4e517c12c2432225e1b4571870716dd"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.298156 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" event={"ID":"9d8d9a60-778d-4318-b05d-7958a905c66d","Type":"ContainerStarted","Data":"77de5e70967d4791f19454a30d1f08b792fa13a9c663dcf2120816ea7d148593"} Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.395840 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.396197 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.896173992 +0000 UTC m=+151.183665661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.398948 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.402569 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:55.902546056 +0000 UTC m=+151.190037725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.501574 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.501973 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.001956773 +0000 UTC m=+151.289448442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.558483 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mhbt7"] Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.614199 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.617231 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.117206212 +0000 UTC m=+151.404697881 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.716699 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.717176 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.217157886 +0000 UTC m=+151.504649555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.732816 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-r5nl7"] Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.754711 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.803219 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf"] Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.818219 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.818622 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.318608426 +0000 UTC m=+151.606100095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:55 crc kubenswrapper[4960]: I1201 19:41:55.919454 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:55 crc kubenswrapper[4960]: E1201 19:41:55.930320 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.430285736 +0000 UTC m=+151.717777405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: W1201 19:41:56.018810 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9750b5a8_c4e9_45be_aac3_8845bff7bc9b.slice/crio-4b1f61b69b77140ff3cfa22b7959651903eeee802e31d630293cbf4d9100db86 WatchSource:0}: Error finding container 4b1f61b69b77140ff3cfa22b7959651903eeee802e31d630293cbf4d9100db86: Status 404 returned error can't find the container with id 4b1f61b69b77140ff3cfa22b7959651903eeee802e31d630293cbf4d9100db86 Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.019693 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z5tw8"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.029646 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.029713 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jgr5w"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.032020 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.032743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.033103 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.533088426 +0000 UTC m=+151.820580095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.033522 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:41:56 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:41:56 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:41:56 crc kubenswrapper[4960]: healthz check failed Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.033572 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:41:56 crc kubenswrapper[4960]: W1201 19:41:56.058416 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54cb7d4f_88bd_4a3b_a87b_3514e9babca6.slice/crio-096b492d14ab4752c373237db3ae3e12e7882b6950715fb133c6dc74940294ea WatchSource:0}: Error finding container 096b492d14ab4752c373237db3ae3e12e7882b6950715fb133c6dc74940294ea: Status 404 returned error can't find the container with id 096b492d14ab4752c373237db3ae3e12e7882b6950715fb133c6dc74940294ea Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.061886 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w95l4"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.121328 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-k8b5p" podStartSLOduration=128.121309325 podStartE2EDuration="2m8.121309325s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.118818669 +0000 UTC m=+151.406310338" watchObservedRunningTime="2025-12-01 19:41:56.121309325 +0000 UTC m=+151.408800994" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.135036 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.135500 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.635473965 +0000 UTC m=+151.922965634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.162005 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.169747 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.194237 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.227773 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z7r8b"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.232276 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.250791 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rcgdg" podStartSLOduration=128.250759225 podStartE2EDuration="2m8.250759225s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.206418778 +0000 UTC m=+151.493910447" watchObservedRunningTime="2025-12-01 19:41:56.250759225 +0000 UTC m=+151.538250894" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.260495 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.266623 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-p7vzg" podStartSLOduration=128.266600205 podStartE2EDuration="2m8.266600205s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.260599113 +0000 UTC m=+151.548090792" watchObservedRunningTime="2025-12-01 19:41:56.266600205 +0000 UTC m=+151.554091944" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.279664 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.280357 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.780340272 +0000 UTC m=+152.067831941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.328571 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.344322 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-dpfnk" podStartSLOduration=128.344296444 podStartE2EDuration="2m8.344296444s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.33954947 +0000 UTC m=+151.627041139" watchObservedRunningTime="2025-12-01 19:41:56.344296444 +0000 UTC m=+151.631788113" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.380755 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.381162 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.881145733 +0000 UTC m=+152.168637402 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.386048 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" event={"ID":"9750b5a8-c4e9-45be-aac3-8845bff7bc9b","Type":"ContainerStarted","Data":"4b1f61b69b77140ff3cfa22b7959651903eeee802e31d630293cbf4d9100db86"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.398290 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" event={"ID":"b8af7379-3edf-45ef-9564-fe4f6cbf8561","Type":"ContainerStarted","Data":"bef783649c99e8b974cdf528d728bda5917ddb32f8f93ffc663449c89dc66cea"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.402041 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-g7thq" event={"ID":"a1cbbec2-92a5-4a94-9932-7ff4f36b6860","Type":"ContainerStarted","Data":"82bd12c491ec0ecf68c593da8dc73a7c45999cfd0e37ebbbf0058e1c30e2310e"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.405771 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.413029 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" event={"ID":"1b41ded7-08c7-4b04-bd13-a286d8a1bca4","Type":"ContainerStarted","Data":"7f5b8420dbcd21ae2ff33c0cdf1dd3c0401ac8201dbb671a95a79a8a0e5be310"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.415293 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" event={"ID":"66243bfe-890f-4184-b937-b26e694adf33","Type":"ContainerStarted","Data":"6043ec70d6f8676b1b2ac1423266e1226396e0daea8b4d6335ab60c4f20e1a46"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.425274 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" event={"ID":"046442a5-907a-4147-8598-f5f836a2e082","Type":"ContainerStarted","Data":"297c073db3cdb81189b3945f8d3f5a92bbed0f9f54fb79aeba4a58b375665e21"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.430908 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9lvnl"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.434050 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" event={"ID":"b18c64d6-cd4c-4d07-9591-4c9868b33b2c","Type":"ContainerStarted","Data":"6dcf544eba3c562dc545dabfb0e77f0a290f8d56ce6f729efc652df24e06d42e"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.436984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" event={"ID":"32c1d373-0f1b-4f75-bf87-eb6927a64389","Type":"ContainerStarted","Data":"ca1d694797f571c285836354c3653a524a81c55b5655d399141a4427bd63ca1f"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.438348 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wv5p9"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.442272 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w95l4" event={"ID":"c30ab298-c528-45be-99af-d8d1135eef2f","Type":"ContainerStarted","Data":"8eb1478c56ea15893897e44e713bbdeed5855496b7d4f47b22b0e6af4ee10876"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.465451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" event={"ID":"952f446b-a8d2-4887-bc67-cddfb9e20791","Type":"ContainerStarted","Data":"e87b72ead306efd690bb568f44d51c0d6abad565863348f2fe7cb3c7a81e5726"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.476174 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.482250 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" event={"ID":"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4","Type":"ContainerStarted","Data":"ca273d6b2e0fde466266ced5827b44703666fc843208f4ff201e871ad23e07cb"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.483157 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.484853 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:56.98483605 +0000 UTC m=+152.272327719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.540552 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j9vkv" event={"ID":"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0","Type":"ContainerStarted","Data":"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.555519 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" event={"ID":"7a8f833e-a2ef-4058-9ba8-99408a789116","Type":"ContainerStarted","Data":"168583ad19ea110ad0a4e0c8d5765e602bb39c4f32b6ff046d9a0d28532eb9ac"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.558257 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" event={"ID":"54cb7d4f-88bd-4a3b-a87b-3514e9babca6","Type":"ContainerStarted","Data":"096b492d14ab4752c373237db3ae3e12e7882b6950715fb133c6dc74940294ea"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.584541 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.585743 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.085724163 +0000 UTC m=+152.373215832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.600069 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jgr5w" event={"ID":"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d","Type":"ContainerStarted","Data":"e4619226f4589dabb199cc79bbb226ae6c13c98a718acca691674f14002a673a"} Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.618362 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" podStartSLOduration=128.618339293 podStartE2EDuration="2m8.618339293s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.526529416 +0000 UTC m=+151.814021085" watchObservedRunningTime="2025-12-01 19:41:56.618339293 +0000 UTC m=+151.905830962" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.687602 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.688205 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.188186193 +0000 UTC m=+152.475677862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.738750 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nz287"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.758481 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-g7thq" Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.812383 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.812497 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.312469666 +0000 UTC m=+152.599961335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.812817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.813191 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.313178568 +0000 UTC m=+152.600670237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.844232 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-c48mv"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.884725 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb"] Dec 01 19:41:56 crc kubenswrapper[4960]: I1201 19:41:56.916059 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:56 crc kubenswrapper[4960]: E1201 19:41:56.916298 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.416263947 +0000 UTC m=+152.703755616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.020689 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.021746 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.521725238 +0000 UTC m=+152.809217087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.031339 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cj28l" podStartSLOduration=129.03131887 podStartE2EDuration="2m9.03131887s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:56.994705928 +0000 UTC m=+152.282197597" watchObservedRunningTime="2025-12-01 19:41:57.03131887 +0000 UTC m=+152.318810539" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.033660 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-g7thq" podStartSLOduration=129.033650781 podStartE2EDuration="2m9.033650781s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.030599548 +0000 UTC m=+152.318091217" watchObservedRunningTime="2025-12-01 19:41:57.033650781 +0000 UTC m=+152.321142450" Dec 01 19:41:57 crc kubenswrapper[4960]: W1201 19:41:57.039448 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c430c72_691c_4a6c_b4c5_9edb65fc8a31.slice/crio-cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df WatchSource:0}: Error finding container cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df: Status 404 returned error can't find the container with id cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.050249 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:41:57 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:41:57 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:41:57 crc kubenswrapper[4960]: healthz check failed Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.050290 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.086901 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ms5tj" podStartSLOduration=129.086877796 podStartE2EDuration="2m9.086877796s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.084460213 +0000 UTC m=+152.371951882" watchObservedRunningTime="2025-12-01 19:41:57.086877796 +0000 UTC m=+152.374369465" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.124733 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.125071 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.625053475 +0000 UTC m=+152.912545144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.128494 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-j9vkv" podStartSLOduration=129.128471059 podStartE2EDuration="2m9.128471059s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.126172049 +0000 UTC m=+152.413663718" watchObservedRunningTime="2025-12-01 19:41:57.128471059 +0000 UTC m=+152.415962728" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.226211 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.226717 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.726696431 +0000 UTC m=+153.014188100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.331934 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.332500 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.832480382 +0000 UTC m=+153.119972051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.433521 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.433872 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:57.93385689 +0000 UTC m=+153.221348549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.535396 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.536326 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.036308289 +0000 UTC m=+153.323799958 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.627063 4960 generic.go:334] "Generic (PLEG): container finished" podID="19bef78c-7725-4036-a57a-7e7335dbaed2" containerID="c474a017ef17b6ef15ac30aff8b12e79ceef743a79d5da411992bcf4fad7b334" exitCode=0 Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.627942 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" event={"ID":"19bef78c-7725-4036-a57a-7e7335dbaed2","Type":"ContainerDied","Data":"c474a017ef17b6ef15ac30aff8b12e79ceef743a79d5da411992bcf4fad7b334"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.630398 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" event={"ID":"0db6e6bb-0673-416b-a48d-4cfa433893bb","Type":"ContainerStarted","Data":"00bf27837416c21eb81914728c154d1f604a81d84bb95270e8b4734f6c7dadbf"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.639676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.640109 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.14009581 +0000 UTC m=+153.427587479 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.642729 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" event={"ID":"66243bfe-890f-4184-b937-b26e694adf33","Type":"ContainerStarted","Data":"a5e4a6e34770c8e4c7da38f82d6e1ad3cadf77c4d57d36b4b8a8caf01237c12a"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.660391 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerStarted","Data":"c802b1029eecf5869c1561db4a2a934a9dc274e5e20550682333f1a2947b1106"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.680199 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8cvhx" event={"ID":"84c9d692-2ab6-42de-b2e9-320ae0561090","Type":"ContainerStarted","Data":"945215ac19fd129e46a1c2b480a101f479c5eaa40fd581caa13fba7a8ac71c68"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.684894 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wv5p9" event={"ID":"ad5cd286-cc5e-4c29-a98c-992994d3ada2","Type":"ContainerStarted","Data":"c4e1816e2055b66301ec151b8ad3ac630015275f907261e6c0a3d6f79321e4e4"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.690069 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" event={"ID":"6a734d24-91a7-4d3c-a719-403b9aa87c9a","Type":"ContainerStarted","Data":"7b7c682fb2bb9ad95e6d4c00c372e462181eeaf2bd09e295fa6c7769898843c5"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.710953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" event={"ID":"314ac1b3-c601-4dda-8374-f76638dbc5cd","Type":"ContainerStarted","Data":"9aededef5ce323c8dc60a1a2ec4bb38f42e8ec68df631e161423a0d9f40e26ac"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.712106 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-r5nl7" podStartSLOduration=129.712082036 podStartE2EDuration="2m9.712082036s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.681335593 +0000 UTC m=+152.968827262" watchObservedRunningTime="2025-12-01 19:41:57.712082036 +0000 UTC m=+152.999573705" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.712286 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8cvhx" podStartSLOduration=6.712280882 podStartE2EDuration="6.712280882s" podCreationTimestamp="2025-12-01 19:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.709443356 +0000 UTC m=+152.996935025" watchObservedRunningTime="2025-12-01 19:41:57.712280882 +0000 UTC m=+152.999772551" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.714085 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" event={"ID":"dd5a0cdf-1451-4787-b68b-98b72c5173d7","Type":"ContainerStarted","Data":"7e2ebbedabaf6226c59f8725cf235d0ac988e980f561bb685fde60e56b433749"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.734043 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" event={"ID":"7ce36e69-5e33-47d3-b526-41844fb4ff8e","Type":"ContainerStarted","Data":"40a6e010265a5023900e29fcbd0f49285decb6253e98193c7282ec8489d3aac3"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.747322 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" event={"ID":"ffd93427-d545-4290-830e-329e8b42dce6","Type":"ContainerStarted","Data":"ff487f4b8a1a5dd3e7b8db229c7f9c24b02df90a6c0a595d53b53d2bf77fb914"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.747408 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.747528 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.247509301 +0000 UTC m=+153.535000970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.747941 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.752554 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.252530413 +0000 UTC m=+153.540022072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.755381 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q2w5q" podStartSLOduration=129.755356919 podStartE2EDuration="2m9.755356919s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.752274856 +0000 UTC m=+153.039766525" watchObservedRunningTime="2025-12-01 19:41:57.755356919 +0000 UTC m=+153.042848588" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.758359 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" event={"ID":"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b","Type":"ContainerStarted","Data":"1d37a0dd47640a30bd1428046b8c0d5888328ab13cd5c41a24d294f1a8f83046"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.764969 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" event={"ID":"cc354d2f-b7ad-4878-8a1d-ffae0d774db3","Type":"ContainerStarted","Data":"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.765282 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.776048 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.780617 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jgr5w" event={"ID":"988d62ee-9dbc-44a4-b67e-6a2baeb2df7d","Type":"ContainerStarted","Data":"046b4c4c812f81e4f8581445574ee8693963e239ee9ec6d89f6ab9edca02ada9"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.796458 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" event={"ID":"de273fd5-2f0f-4532-b6ca-67cd9893ab26","Type":"ContainerStarted","Data":"ef3c7a0853ec80ff7e8751cbaf2de9eedd66c45d2460f67009c4169b52d363d9"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.797292 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.798055 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" podStartSLOduration=129.798036745 podStartE2EDuration="2m9.798036745s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.794174247 +0000 UTC m=+153.081665916" watchObservedRunningTime="2025-12-01 19:41:57.798036745 +0000 UTC m=+153.085528404" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.803720 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" event={"ID":"ff6d1596-2e90-47bb-a6ff-4335b49328f6","Type":"ContainerStarted","Data":"de6054faf4c80cf0f6342fc21981df5c75bf68403aa4fe6e10051f66fc09d872"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.821857 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" event={"ID":"9b8a78a1-d86c-498b-8bd2-52ff3687455e","Type":"ContainerStarted","Data":"a74d4cd8df0fea7914b0d3f017c156f43cfe1988da6c8b0cf447d18ff647313d"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.823650 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.831603 4960 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-sgctq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.831660 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" podUID="9b8a78a1-d86c-498b-8bd2-52ff3687455e" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.838176 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" event={"ID":"3c430c72-691c-4a6c-b4c5-9edb65fc8a31","Type":"ContainerStarted","Data":"cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.848418 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" event={"ID":"a3386a2b-dc67-4f1d-ba89-04e29243819b","Type":"ContainerStarted","Data":"d52de6b4ec3ccb93adf63d26922484b55a09a26d961ffb81cfa6b5c1eb539bd8"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.854427 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.854692 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jgr5w" podStartSLOduration=6.854673405 podStartE2EDuration="6.854673405s" podCreationTimestamp="2025-12-01 19:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.853497589 +0000 UTC m=+153.140989268" watchObservedRunningTime="2025-12-01 19:41:57.854673405 +0000 UTC m=+153.142165074" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.855191 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.856991 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.356966054 +0000 UTC m=+153.644457723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.857884 4960 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-qx5vs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" start-of-body= Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.857946 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" podUID="7a8f833e-a2ef-4058-9ba8-99408a789116" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.26:5443/healthz\": dial tcp 10.217.0.26:5443: connect: connection refused" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.877316 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" event={"ID":"b18c64d6-cd4c-4d07-9591-4c9868b33b2c","Type":"ContainerStarted","Data":"ffa0e5e21494adca1a99a8144aa6e95e763e0ed58ef7f35f936b0c7886c7844f"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.879258 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" podStartSLOduration=129.8792405 podStartE2EDuration="2m9.8792405s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.876326792 +0000 UTC m=+153.163818461" watchObservedRunningTime="2025-12-01 19:41:57.8792405 +0000 UTC m=+153.166732169" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.901200 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" podStartSLOduration=129.901175446 podStartE2EDuration="2m9.901175446s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.901155135 +0000 UTC m=+153.188646804" watchObservedRunningTime="2025-12-01 19:41:57.901175446 +0000 UTC m=+153.188667115" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.911902 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" event={"ID":"9750b5a8-c4e9-45be-aac3-8845bff7bc9b","Type":"ContainerStarted","Data":"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.912649 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.924837 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.941831 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" podStartSLOduration=129.94180942 podStartE2EDuration="2m9.94180942s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.94149647 +0000 UTC m=+153.228988139" watchObservedRunningTime="2025-12-01 19:41:57.94180942 +0000 UTC m=+153.229301089" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.947142 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" event={"ID":"1b41ded7-08c7-4b04-bd13-a286d8a1bca4","Type":"ContainerStarted","Data":"9c151fb2444b07cd44d9bfdc26c529a9f5b784a838dc5fd480c88b60912fe5c8"} Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.958561 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.975391 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" podStartSLOduration=129.975365649 podStartE2EDuration="2m9.975365649s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:57.973658466 +0000 UTC m=+153.261150155" watchObservedRunningTime="2025-12-01 19:41:57.975365649 +0000 UTC m=+153.262857318" Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.976016 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" event={"ID":"9f6b417f-45c1-48e6-abac-6b4ea9ddf1b4","Type":"ContainerStarted","Data":"fbd1dd46c0686c9eb3e2f9e0b8596689f364a11abb443c7c1cee56ee136aa672"} Dec 01 19:41:57 crc kubenswrapper[4960]: E1201 19:41:57.983589 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.483506646 +0000 UTC m=+153.770998315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:57 crc kubenswrapper[4960]: I1201 19:41:57.992810 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" event={"ID":"116b4f38-75e4-4b24-8e46-41e9d78a8c81","Type":"ContainerStarted","Data":"49e58e459ddf78b76c599e5c45d1c0bfaffabb73a3e16b2b8862155c0d39379e"} Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.018909 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" podStartSLOduration=130.018882379 podStartE2EDuration="2m10.018882379s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:58.013158925 +0000 UTC m=+153.300650604" watchObservedRunningTime="2025-12-01 19:41:58.018882379 +0000 UTC m=+153.306374038" Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.038734 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:41:58 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:41:58 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:41:58 crc kubenswrapper[4960]: healthz check failed Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.038970 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.053945 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-69vvk" podStartSLOduration=130.053917583 podStartE2EDuration="2m10.053917583s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:58.039136544 +0000 UTC m=+153.326628203" watchObservedRunningTime="2025-12-01 19:41:58.053917583 +0000 UTC m=+153.341409252" Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.068219 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.071028 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.570991831 +0000 UTC m=+153.858483500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.082732 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-mhrdh" podStartSLOduration=130.082712767 podStartE2EDuration="2m10.082712767s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:58.078685625 +0000 UTC m=+153.366177294" watchObservedRunningTime="2025-12-01 19:41:58.082712767 +0000 UTC m=+153.370204436" Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.122688 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" podStartSLOduration=130.122642969 podStartE2EDuration="2m10.122642969s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:58.121911107 +0000 UTC m=+153.409402776" watchObservedRunningTime="2025-12-01 19:41:58.122642969 +0000 UTC m=+153.410134638" Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.176023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.176417 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.676400821 +0000 UTC m=+153.963892490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.277697 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.278462 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.778437438 +0000 UTC m=+154.065929107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.379362 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.380213 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.880185607 +0000 UTC m=+154.167677276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.482033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.482341 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.982313518 +0000 UTC m=+154.269805187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.482390 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.482759 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:58.982752271 +0000 UTC m=+154.270243940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.583867 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.584331 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.084310084 +0000 UTC m=+154.371801753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.690172 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.690911 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.19089745 +0000 UTC m=+154.478389119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.791232 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.791408 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.29137733 +0000 UTC m=+154.578868999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.791535 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.791878 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.291863365 +0000 UTC m=+154.579355034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.892651 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.892872 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.39285557 +0000 UTC m=+154.680347239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.893232 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.893559 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.393551352 +0000 UTC m=+154.681043011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:58 crc kubenswrapper[4960]: I1201 19:41:58.994223 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:58 crc kubenswrapper[4960]: E1201 19:41:58.994660 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.49464325 +0000 UTC m=+154.782134919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.004271 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" event={"ID":"dd5a0cdf-1451-4787-b68b-98b72c5173d7","Type":"ContainerStarted","Data":"010a0d11d7cdcbfdacc7fbb5808153df49c3c99c1c71dd3bf57624b7a3105260"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.004435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" event={"ID":"dd5a0cdf-1451-4787-b68b-98b72c5173d7","Type":"ContainerStarted","Data":"3460d9513920671403f677ae139d258ec16c0c3fa27ebf6fc3bfc224a0fecfae"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.006198 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" event={"ID":"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b","Type":"ContainerStarted","Data":"0cd22dbb6570fbbced893fc2c839aba557126c8632a775200658ebc3f6f85f1e"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.009818 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" event={"ID":"a3386a2b-dc67-4f1d-ba89-04e29243819b","Type":"ContainerStarted","Data":"d52018733532cc27ee7b1b917107867af93ffcbc8eac4da2c70eb2be77b7ad57"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.015642 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" event={"ID":"6a734d24-91a7-4d3c-a719-403b9aa87c9a","Type":"ContainerStarted","Data":"b5f3787253eb05fbfebd04e8c040947367ff1e56420704bd2c2722167694e799"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.017770 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" event={"ID":"b8af7379-3edf-45ef-9564-fe4f6cbf8561","Type":"ContainerStarted","Data":"8eb2941be99a159c308877439ae53beaa8c9e232434f4f1a46e1c0a83b8a396a"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.023049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w95l4" event={"ID":"c30ab298-c528-45be-99af-d8d1135eef2f","Type":"ContainerStarted","Data":"ee93251d24843fce59a506e1c8552d801a739a858343084839f04276f607d3fe"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.023728 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.024730 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-w95l4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.024781 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w95l4" podUID="c30ab298-c528-45be-99af-d8d1135eef2f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.029386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" event={"ID":"ffd93427-d545-4290-830e-329e8b42dce6","Type":"ContainerStarted","Data":"59362a5c5fd72459380f999b09187df3eb3a661678ab673da47ad6a43f9caf58"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.029417 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.029429 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" event={"ID":"ffd93427-d545-4290-830e-329e8b42dce6","Type":"ContainerStarted","Data":"f04977190354213db99ec7644a410f759b4afef7064e1e036a4e5bd42ff56a92"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.031919 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:41:59 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:41:59 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:41:59 crc kubenswrapper[4960]: healthz check failed Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.032194 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.036306 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" event={"ID":"3c430c72-691c-4a6c-b4c5-9edb65fc8a31","Type":"ContainerStarted","Data":"fb65c6f74e217e163c05263be53dfb9dd14f26f56b08f9429f1b94ee5c2cc8a8"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.041777 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-z5tw8" podStartSLOduration=131.041763021 podStartE2EDuration="2m11.041763021s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.041421461 +0000 UTC m=+154.328913130" watchObservedRunningTime="2025-12-01 19:41:59.041763021 +0000 UTC m=+154.329254690" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.048465 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" event={"ID":"7a8f833e-a2ef-4058-9ba8-99408a789116","Type":"ContainerStarted","Data":"9039bbc39b50ec7e0eb0ac615290ec4ec180cb27a71665b04922fdea37776b64"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.051168 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" event={"ID":"314ac1b3-c601-4dda-8374-f76638dbc5cd","Type":"ContainerStarted","Data":"da07021504d7a28929b82fa0fdba31c97ab5b0d78684f997f84bb3dbc2550216"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.051681 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.055741 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" event={"ID":"ff6d1596-2e90-47bb-a6ff-4335b49328f6","Type":"ContainerStarted","Data":"954765a299d150670f08cc89d89047d6e113b99c1ccd1f0a8cfd288096f717b9"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.058061 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-qx5vs" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.058428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" event={"ID":"9b8a78a1-d86c-498b-8bd2-52ff3687455e","Type":"ContainerStarted","Data":"2aa731ef09a5dae8906c7927c1091464464a24f888b22e24f76fa9b7918a481e"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.058529 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.066365 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" event={"ID":"116b4f38-75e4-4b24-8e46-41e9d78a8c81","Type":"ContainerStarted","Data":"19b8b774ab435fbf6cd70dd6443e54c3b1b5162d81b3eaf07323f53b10585782"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.066420 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" event={"ID":"116b4f38-75e4-4b24-8e46-41e9d78a8c81","Type":"ContainerStarted","Data":"7a6118feb4074d677369ec6bd34383eaf1d924554e08e8539a37bcc99139d863"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.081677 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sgctq" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.088498 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" event={"ID":"bf0fccd8-8b4e-422e-abdc-ecf769334f1d","Type":"ContainerStarted","Data":"318d73e3dc98c27a6f869b81f8411885545f921479a6ce67d2a8b3afc9866984"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.095831 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-6xgg5" podStartSLOduration=131.095792961 podStartE2EDuration="2m11.095792961s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.090213771 +0000 UTC m=+154.377705440" watchObservedRunningTime="2025-12-01 19:41:59.095792961 +0000 UTC m=+154.383284630" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.096477 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.100480 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.600466233 +0000 UTC m=+154.887957902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.128377 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wv5p9" event={"ID":"ad5cd286-cc5e-4c29-a98c-992994d3ada2","Type":"ContainerStarted","Data":"6ba979f549066d5ec6841aba1b37346be074302c59db2bd0638f3f22909c6235"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.128435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wv5p9" event={"ID":"ad5cd286-cc5e-4c29-a98c-992994d3ada2","Type":"ContainerStarted","Data":"386b4e044a29fd74839de8ae266da9ed8e0db14367f0d3dbae672e6fdfd988ad"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.129142 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wv5p9" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.142382 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mhbt7" event={"ID":"b18c64d6-cd4c-4d07-9591-4c9868b33b2c","Type":"ContainerStarted","Data":"3f9cb3cd292fbabe204eb1acddd2aecc37a40509ae2dfbfb709dd4cea7e301c3"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.145368 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" event={"ID":"0db6e6bb-0673-416b-a48d-4cfa433893bb","Type":"ContainerStarted","Data":"cab0e5ed2785c3150bedc845fecd860e33264ee5d67b49d02acb9483f262b391"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.147039 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" event={"ID":"54cb7d4f-88bd-4a3b-a87b-3514e9babca6","Type":"ContainerStarted","Data":"a07a253c0f3fb51e7c44d1d5194b82e6f19a07702f0cf53330e310f7f0bc4130"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.151629 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qhvh" event={"ID":"952f446b-a8d2-4887-bc67-cddfb9e20791","Type":"ContainerStarted","Data":"04dda8872b0727c6d90c194c733105978e5679f128b217deb7d7ff504bf617b3"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.161546 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" event={"ID":"19bef78c-7725-4036-a57a-7e7335dbaed2","Type":"ContainerStarted","Data":"1977c72b0259c03b35d8e13c6e522bd87baf6caeb48395cd631a2cefe365d0bf"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.161703 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" event={"ID":"19bef78c-7725-4036-a57a-7e7335dbaed2","Type":"ContainerStarted","Data":"36e8d91aeaf4ed0d8ce7f988adae814f6acc018391ca9113b4df6ad3fddf9494"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.205436 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" event={"ID":"cf21cb96-ba0f-4872-80bc-3f09db8f8ca8","Type":"ContainerStarted","Data":"511ea9feeabb68a90253771c8e83f5de04893cfa9b2bcfe264605c6eb3040890"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.205819 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.207025 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.707006367 +0000 UTC m=+154.994498026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.225759 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.226254 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.227152 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-66whs" podStartSLOduration=131.227107107 podStartE2EDuration="2m11.227107107s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.169306423 +0000 UTC m=+154.456798092" watchObservedRunningTime="2025-12-01 19:41:59.227107107 +0000 UTC m=+154.514598776" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.239388 4960 generic.go:334] "Generic (PLEG): container finished" podID="21ee4d43-3a2b-48a0-a5dd-332abfab285e" containerID="62b0e1e84c5332db91881b4878ac6cbe5ce8af532934f1f0a899a05d60b0c9fb" exitCode=0 Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.239474 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" event={"ID":"21ee4d43-3a2b-48a0-a5dd-332abfab285e","Type":"ContainerDied","Data":"62b0e1e84c5332db91881b4878ac6cbe5ce8af532934f1f0a899a05d60b0c9fb"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.251483 4960 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5g7zf container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.13:8443/livez\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.251587 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" podUID="19bef78c-7725-4036-a57a-7e7335dbaed2" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.13:8443/livez\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.265522 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerStarted","Data":"90d2540886ef78985b81521d66781c786eb822df3aa3ec28b19ea93c432d7624"} Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.265579 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.280744 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w6wcz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.280833 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.291780 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-6hdx9" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.297675 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" podStartSLOduration=131.297651848 podStartE2EDuration="2m11.297651848s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.225513899 +0000 UTC m=+154.513005568" watchObservedRunningTime="2025-12-01 19:41:59.297651848 +0000 UTC m=+154.585143517" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.317277 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.325078 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.82505949 +0000 UTC m=+155.112551159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.369803 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-sll8n" podStartSLOduration=131.369777849 podStartE2EDuration="2m11.369777849s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.295343149 +0000 UTC m=+154.582834818" watchObservedRunningTime="2025-12-01 19:41:59.369777849 +0000 UTC m=+154.657269518" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.382432 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-w95l4" podStartSLOduration=131.382403002 podStartE2EDuration="2m11.382403002s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.370409507 +0000 UTC m=+154.657901176" watchObservedRunningTime="2025-12-01 19:41:59.382403002 +0000 UTC m=+154.669894671" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.418937 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.420833 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:41:59.920801987 +0000 UTC m=+155.208293656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.437916 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" podStartSLOduration=131.437896766 podStartE2EDuration="2m11.437896766s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.434774241 +0000 UTC m=+154.722265910" watchObservedRunningTime="2025-12-01 19:41:59.437896766 +0000 UTC m=+154.725388435" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.475419 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-2xqzt" podStartSLOduration=131.475394835 podStartE2EDuration="2m11.475394835s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.475367814 +0000 UTC m=+154.762859483" watchObservedRunningTime="2025-12-01 19:41:59.475394835 +0000 UTC m=+154.762886504" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.521052 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.521508 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.021494884 +0000 UTC m=+155.308986553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.536781 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qdx9p" podStartSLOduration=131.536758957 podStartE2EDuration="2m11.536758957s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.529893559 +0000 UTC m=+154.817385238" watchObservedRunningTime="2025-12-01 19:41:59.536758957 +0000 UTC m=+154.824250626" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.625016 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.625347 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.125326516 +0000 UTC m=+155.412818185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.667855 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8p4zf" podStartSLOduration=131.667828426 podStartE2EDuration="2m11.667828426s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.666796945 +0000 UTC m=+154.954288614" watchObservedRunningTime="2025-12-01 19:41:59.667828426 +0000 UTC m=+154.955320095" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.716294 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-c48mv" podStartSLOduration=131.716253337 podStartE2EDuration="2m11.716253337s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.715754251 +0000 UTC m=+155.003245920" watchObservedRunningTime="2025-12-01 19:41:59.716253337 +0000 UTC m=+155.003745006" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.733071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.733439 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.233425258 +0000 UTC m=+155.520916927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.784352 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nz287" podStartSLOduration=131.784328982 podStartE2EDuration="2m11.784328982s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.753003632 +0000 UTC m=+155.040495301" watchObservedRunningTime="2025-12-01 19:41:59.784328982 +0000 UTC m=+155.071820651" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.834231 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.834812 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.334790514 +0000 UTC m=+155.622282183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.866420 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" podStartSLOduration=131.866402414 podStartE2EDuration="2m11.866402414s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.785856549 +0000 UTC m=+155.073348218" watchObservedRunningTime="2025-12-01 19:41:59.866402414 +0000 UTC m=+155.153894083" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.867407 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5vwj5" podStartSLOduration=131.867402615 podStartE2EDuration="2m11.867402615s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.864951011 +0000 UTC m=+155.152442680" watchObservedRunningTime="2025-12-01 19:41:59.867402615 +0000 UTC m=+155.154894284" Dec 01 19:41:59 crc kubenswrapper[4960]: I1201 19:41:59.950369 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:41:59 crc kubenswrapper[4960]: E1201 19:41:59.950804 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.450790756 +0000 UTC m=+155.738282425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.032575 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wv5p9" podStartSLOduration=9.032549178 podStartE2EDuration="9.032549178s" podCreationTimestamp="2025-12-01 19:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:41:59.953928411 +0000 UTC m=+155.241420300" watchObservedRunningTime="2025-12-01 19:42:00.032549178 +0000 UTC m=+155.320040847" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.034984 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:00 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:00 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:00 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.035067 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.051372 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.051870 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.551833613 +0000 UTC m=+155.839325282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.071487 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" podStartSLOduration=132.07146623 podStartE2EDuration="2m12.07146623s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:00.040893241 +0000 UTC m=+155.328384910" watchObservedRunningTime="2025-12-01 19:42:00.07146623 +0000 UTC m=+155.358957899" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.143515 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9lvnl" podStartSLOduration=132.143494426 podStartE2EDuration="2m12.143494426s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:00.140808374 +0000 UTC m=+155.428300053" watchObservedRunningTime="2025-12-01 19:42:00.143494426 +0000 UTC m=+155.430986095" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.153383 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.153780 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.653765647 +0000 UTC m=+155.941257316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.254354 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.254565 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.754528186 +0000 UTC m=+156.042019855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.254763 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.255173 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.755164516 +0000 UTC m=+156.042656185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.281301 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" event={"ID":"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b","Type":"ContainerStarted","Data":"3c23cae94486d4868cff192cb0374a9e4e1c4fe3938e3047be86d901705ac442"} Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.281375 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" event={"ID":"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b","Type":"ContainerStarted","Data":"cf81b2db7de1c95b997e1940ef4321bb04aec99ab3811d38dd3fb5280015ff25"} Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.287755 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" event={"ID":"21ee4d43-3a2b-48a0-a5dd-332abfab285e","Type":"ContainerStarted","Data":"272255b28e4057963060d4d94d0ca612e786fd45c0cafc35698615bd2327bc86"} Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.288788 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-w95l4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.288816 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w6wcz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.288849 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w95l4" podUID="c30ab298-c528-45be-99af-d8d1135eef2f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.288871 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.338033 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" podStartSLOduration=132.338013571 podStartE2EDuration="2m12.338013571s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:00.336527836 +0000 UTC m=+155.624019515" watchObservedRunningTime="2025-12-01 19:42:00.338013571 +0000 UTC m=+155.625505230" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.356940 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.358216 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.858184373 +0000 UTC m=+156.145676042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.460193 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.460640 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:00.960619713 +0000 UTC m=+156.248111382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.563081 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.563347 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.06331339 +0000 UTC m=+156.350805059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.563897 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.564471 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.064456325 +0000 UTC m=+156.351947994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.603692 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.605543 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: W1201 19:42:00.623496 4960 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.623544 4960 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.639660 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.670025 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.670426 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.670494 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.670518 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtblk\" (UniqueName: \"kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.670644 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.170627148 +0000 UTC m=+156.458118817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.735859 4960 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.751367 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.752409 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.758619 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.771954 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.772013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.772068 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.772087 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtblk\" (UniqueName: \"kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.772375 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.272355697 +0000 UTC m=+156.559847366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.772856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.772922 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.779355 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.830928 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtblk\" (UniqueName: \"kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk\") pod \"certified-operators-jjmnx\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.873206 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.873427 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.373396513 +0000 UTC m=+156.660888182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.873587 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.873649 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.873782 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.873844 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2lqt\" (UniqueName: \"kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.874273 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.374251609 +0000 UTC m=+156.661743278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.975104 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.975346 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.475317398 +0000 UTC m=+156.762809067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.975403 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.975509 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.975555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2lqt\" (UniqueName: \"kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.975665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: E1201 19:42:00.976323 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.476306588 +0000 UTC m=+156.763798257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.976831 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.976913 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.982237 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:00 crc kubenswrapper[4960]: I1201 19:42:00.983478 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.025920 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2lqt\" (UniqueName: \"kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt\") pod \"community-operators-zlm2j\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.030433 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:01 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:01 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:01 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.030681 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.044921 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.077456 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.077680 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.577646793 +0000 UTC m=+156.865138462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.077741 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.077842 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jmxm\" (UniqueName: \"kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.077904 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.077927 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.078264 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.578246752 +0000 UTC m=+156.865738421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.091572 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.116321 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.119621 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.149564 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.179822 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180043 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180072 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180136 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jmxm\" (UniqueName: \"kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180162 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180179 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.180240 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm2vq\" (UniqueName: \"kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.180361 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.680341022 +0000 UTC m=+156.967832691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.181019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.181469 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.208257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jmxm\" (UniqueName: \"kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm\") pod \"certified-operators-lmk2r\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.284513 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.284566 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.284603 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.284677 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm2vq\" (UniqueName: \"kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.285541 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.285674 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.285726 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.78570569 +0000 UTC m=+157.073197359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.298824 4960 generic.go:334] "Generic (PLEG): container finished" podID="3c430c72-691c-4a6c-b4c5-9edb65fc8a31" containerID="fb65c6f74e217e163c05263be53dfb9dd14f26f56b08f9429f1b94ee5c2cc8a8" exitCode=0 Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.298926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" event={"ID":"3c430c72-691c-4a6c-b4c5-9edb65fc8a31","Type":"ContainerDied","Data":"fb65c6f74e217e163c05263be53dfb9dd14f26f56b08f9429f1b94ee5c2cc8a8"} Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.378847 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" event={"ID":"4ea4ed8d-6a33-4864-8fd8-67d1d617bf0b","Type":"ContainerStarted","Data":"b6038216efc6e5d65f31399b9c99ed2aca70b7a768a9adff23188a834840fac0"} Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.379513 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm2vq\" (UniqueName: \"kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq\") pod \"community-operators-lfdlz\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.381377 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-w95l4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.381456 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w95l4" podUID="c30ab298-c528-45be-99af-d8d1135eef2f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.387380 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.387784 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.887767969 +0000 UTC m=+157.175259638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.485659 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.491605 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.499825 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:01.999797999 +0000 UTC m=+157.287289668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.525662 4960 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-01T19:42:00.73589694Z","Handler":null,"Name":""} Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.594842 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.595041 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-01 19:42:02.09501133 +0000 UTC m=+157.382502999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.595291 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: E1201 19:42:01.595679 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-01 19:42:02.095672939 +0000 UTC m=+157.383164608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ccbrv" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.646707 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-z7r8b" podStartSLOduration=10.646681978 podStartE2EDuration="10.646681978s" podCreationTimestamp="2025-12-01 19:41:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:01.446741249 +0000 UTC m=+156.734232918" watchObservedRunningTime="2025-12-01 19:42:01.646681978 +0000 UTC m=+156.934173647" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.647278 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.658414 4960 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.658462 4960 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.700699 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.714650 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.805883 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.812204 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.812266 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.911570 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ccbrv\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.928801 4960 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-jjmnx" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.928870 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.978791 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:01 crc kubenswrapper[4960]: I1201 19:42:01.998007 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.033551 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:02 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:02 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:02 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.033631 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.033654 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.039590 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.190860 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.286462 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.384471 4960 generic.go:334] "Generic (PLEG): container finished" podID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerID="a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05" exitCode=0 Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.384567 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerDied","Data":"a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.384631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerStarted","Data":"819bb0b87e374111614b6b8fbd9ece498dd95f4144025034a419be7af43225cf"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.386638 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerStarted","Data":"13d0a8feebb80bac2dd4d8c8964b67afcd4006c68d165d5399796e7cc8565cce"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.386863 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.389380 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerID="c5c4d2b87c939f6933da8f4e03b843e19dd2f3f957949bb31491d39929a81401" exitCode=0 Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.389532 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerDied","Data":"c5c4d2b87c939f6933da8f4e03b843e19dd2f3f957949bb31491d39929a81401"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.389585 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerStarted","Data":"237af35b8b1800aa903849d6825632d5db942e39aeb79828176f80856d003af9"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.392022 4960 generic.go:334] "Generic (PLEG): container finished" podID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerID="041dafbd13346d3fde0f7142edebb895870e1ff9d757798c345cd6875c8415b3" exitCode=0 Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.392944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerDied","Data":"041dafbd13346d3fde0f7142edebb895870e1ff9d757798c345cd6875c8415b3"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.392963 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerStarted","Data":"b0d3eef713b4fe4ae19b780413a34a14e98e30cef7a4d456c0c5538579f3822f"} Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.470138 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.639980 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.703819 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:42:02 crc kubenswrapper[4960]: E1201 19:42:02.704198 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c430c72-691c-4a6c-b4c5-9edb65fc8a31" containerName="collect-profiles" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.704217 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c430c72-691c-4a6c-b4c5-9edb65fc8a31" containerName="collect-profiles" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.704368 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c430c72-691c-4a6c-b4c5-9edb65fc8a31" containerName="collect-profiles" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.705672 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.708927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.716681 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.716802 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume\") pod \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.716890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsqrt\" (UniqueName: \"kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt\") pod \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.717033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume\") pod \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\" (UID: \"3c430c72-691c-4a6c-b4c5-9edb65fc8a31\") " Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.720632 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume" (OuterVolumeSpecName: "config-volume") pod "3c430c72-691c-4a6c-b4c5-9edb65fc8a31" (UID: "3c430c72-691c-4a6c-b4c5-9edb65fc8a31"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.725605 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt" (OuterVolumeSpecName: "kube-api-access-qsqrt") pod "3c430c72-691c-4a6c-b4c5-9edb65fc8a31" (UID: "3c430c72-691c-4a6c-b4c5-9edb65fc8a31"). InnerVolumeSpecName "kube-api-access-qsqrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.725649 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3c430c72-691c-4a6c-b4c5-9edb65fc8a31" (UID: "3c430c72-691c-4a6c-b4c5-9edb65fc8a31"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818629 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lmv2\" (UniqueName: \"kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818656 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818780 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818795 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.818806 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsqrt\" (UniqueName: \"kubernetes.io/projected/3c430c72-691c-4a6c-b4c5-9edb65fc8a31-kube-api-access-qsqrt\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.920346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.920457 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lmv2\" (UniqueName: \"kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.920486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.921092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.922436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:02 crc kubenswrapper[4960]: I1201 19:42:02.954506 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lmv2\" (UniqueName: \"kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2\") pod \"redhat-marketplace-l7ldv\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.028194 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.031581 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:03 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:03 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:03 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.032015 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.113795 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.115224 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.127450 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.180096 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.181216 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.183922 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.184232 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.189356 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.227281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.227374 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.227468 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v6mq\" (UniqueName: \"kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.282126 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:42:03 crc kubenswrapper[4960]: W1201 19:42:03.291006 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65475df3_6ce0_44d3_a5ed_861e34334058.slice/crio-1ff3fb068b0fb8bc4087ede0409e138a23470af64d25d507c6e49db6ee8bdfd6 WatchSource:0}: Error finding container 1ff3fb068b0fb8bc4087ede0409e138a23470af64d25d507c6e49db6ee8bdfd6: Status 404 returned error can't find the container with id 1ff3fb068b0fb8bc4087ede0409e138a23470af64d25d507c6e49db6ee8bdfd6 Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.329297 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.329416 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.329482 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.329530 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.329617 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v6mq\" (UniqueName: \"kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.330828 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.330904 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.345189 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.383268 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v6mq\" (UniqueName: \"kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq\") pod \"redhat-marketplace-kmczg\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.428609 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerStarted","Data":"1ff3fb068b0fb8bc4087ede0409e138a23470af64d25d507c6e49db6ee8bdfd6"} Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.437737 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.439302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.439379 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.439572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.457243 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" event={"ID":"41dd0037-d130-4fcf-8507-4dfa3059a9d9","Type":"ContainerStarted","Data":"faedc782190686d1686a51a5032b561d0ac9c718712d35b4a33c5115b051f82a"} Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.457311 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" event={"ID":"41dd0037-d130-4fcf-8507-4dfa3059a9d9","Type":"ContainerStarted","Data":"42097c77da0911afbf7d83a8a1db301627769d32b84eac5b547f87ea687d4e33"} Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.458622 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.470073 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.490659 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" event={"ID":"3c430c72-691c-4a6c-b4c5-9edb65fc8a31","Type":"ContainerDied","Data":"cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df"} Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.490707 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc38a6d7861a5827412f524939038874079b80bba1db5af578edd4a29bb1a9df" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.490781 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.491408 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" podStartSLOduration=135.491398108 podStartE2EDuration="2m15.491398108s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:03.489722387 +0000 UTC m=+158.777214056" watchObservedRunningTime="2025-12-01 19:42:03.491398108 +0000 UTC m=+158.778889777" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.493046 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerID="6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c" exitCode=0 Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.493088 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerDied","Data":"6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c"} Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.502088 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.690903 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.706886 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.708104 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.712021 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.716070 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.845449 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.845529 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4wvs\" (UniqueName: \"kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.845616 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.899033 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 01 19:42:03 crc kubenswrapper[4960]: W1201 19:42:03.915343 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podface8246_9a4b_4267_b919_2f78a451ba18.slice/crio-1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1 WatchSource:0}: Error finding container 1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1: Status 404 returned error can't find the container with id 1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1 Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.946901 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.947005 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4wvs\" (UniqueName: \"kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.947067 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.947783 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.947899 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.973994 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.974644 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.985188 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4wvs\" (UniqueName: \"kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs\") pod \"redhat-operators-fwhqw\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.989158 4960 patch_prober.go:28] interesting pod/console-f9d7485db-j9vkv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.33:8443/health\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 01 19:42:03 crc kubenswrapper[4960]: I1201 19:42:03.989250 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-j9vkv" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.33:8443/health\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.027294 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.028688 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.031932 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:04 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:04 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:04 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.032012 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.109784 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.110941 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.122193 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.234776 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.242379 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5g7zf" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.253931 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.253973 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdsm5\" (UniqueName: \"kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.254006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.256468 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:42:04 crc kubenswrapper[4960]: W1201 19:42:04.272519 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c55929c_b18d_4719_944e_54176bdb96bb.slice/crio-244d51c5f070ed3060757729c980afa0eb2e8bb66f8c1e3201daad725cd3ca0f WatchSource:0}: Error finding container 244d51c5f070ed3060757729c980afa0eb2e8bb66f8c1e3201daad725cd3ca0f: Status 404 returned error can't find the container with id 244d51c5f070ed3060757729c980afa0eb2e8bb66f8c1e3201daad725cd3ca0f Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.283483 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.300627 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.300697 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.338909 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.357178 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.357226 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdsm5\" (UniqueName: \"kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.357302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.360471 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.364743 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.442139 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdsm5\" (UniqueName: \"kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5\") pod \"redhat-operators-m4x5q\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.457797 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.535573 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerStarted","Data":"244d51c5f070ed3060757729c980afa0eb2e8bb66f8c1e3201daad725cd3ca0f"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.538542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"face8246-9a4b-4267-b919-2f78a451ba18","Type":"ContainerStarted","Data":"1cb851ef1d549cd8ed68433c2123953be674b69744ebe0476a72e95c811bd749"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.538614 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"face8246-9a4b-4267-b919-2f78a451ba18","Type":"ContainerStarted","Data":"1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.558851 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.5588297519999998 podStartE2EDuration="1.558829752s" podCreationTimestamp="2025-12-01 19:42:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:04.554731507 +0000 UTC m=+159.842223176" watchObservedRunningTime="2025-12-01 19:42:04.558829752 +0000 UTC m=+159.846321421" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.587562 4960 generic.go:334] "Generic (PLEG): container finished" podID="65475df3-6ce0-44d3-a5ed-861e34334058" containerID="cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702" exitCode=0 Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.587702 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerDied","Data":"cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.605268 4960 generic.go:334] "Generic (PLEG): container finished" podID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerID="59aae7d51edbaed8f392ccc805f9a7f05a9fb9c560510832d35cde707bf0d2c0" exitCode=0 Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.605382 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerDied","Data":"59aae7d51edbaed8f392ccc805f9a7f05a9fb9c560510832d35cde707bf0d2c0"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.605454 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerStarted","Data":"70a53a1c4fcb710de2cc018a28c714b3df7fca065ef6a21e29fcfe393db82b77"} Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.614221 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-m5ktp" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.718295 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-w95l4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.718813 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w95l4" podUID="c30ab298-c528-45be-99af-d8d1135eef2f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.718507 4960 patch_prober.go:28] interesting pod/downloads-7954f5f757-w95l4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.719290 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w95l4" podUID="c30ab298-c528-45be-99af-d8d1135eef2f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.39:8080/\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.720015 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:42:04 crc kubenswrapper[4960]: I1201 19:42:04.885830 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:42:04 crc kubenswrapper[4960]: W1201 19:42:04.905990 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdba2587_565b_4ffc_9448_5b757cc58410.slice/crio-0dc878703ab0491ab2a17ce1207030a426a9f261f799e371ef987877f6e669c4 WatchSource:0}: Error finding container 0dc878703ab0491ab2a17ce1207030a426a9f261f799e371ef987877f6e669c4: Status 404 returned error can't find the container with id 0dc878703ab0491ab2a17ce1207030a426a9f261f799e371ef987877f6e669c4 Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.030797 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:05 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:05 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:05 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.030898 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.617324 4960 generic.go:334] "Generic (PLEG): container finished" podID="cdba2587-565b-4ffc-9448-5b757cc58410" containerID="2e519b96a259b4b2837e9eca5df3feceb9a05ea5780112f84a53702c6eb4b427" exitCode=0 Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.617515 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerDied","Data":"2e519b96a259b4b2837e9eca5df3feceb9a05ea5780112f84a53702c6eb4b427"} Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.617837 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerStarted","Data":"0dc878703ab0491ab2a17ce1207030a426a9f261f799e371ef987877f6e669c4"} Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.625787 4960 generic.go:334] "Generic (PLEG): container finished" podID="3c55929c-b18d-4719-944e-54176bdb96bb" containerID="146dec8baf84c5bf60700f546db42ca1c022600e6abb0af144e426bf2d7affc3" exitCode=0 Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.625960 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerDied","Data":"146dec8baf84c5bf60700f546db42ca1c022600e6abb0af144e426bf2d7affc3"} Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.629282 4960 generic.go:334] "Generic (PLEG): container finished" podID="face8246-9a4b-4267-b919-2f78a451ba18" containerID="1cb851ef1d549cd8ed68433c2123953be674b69744ebe0476a72e95c811bd749" exitCode=0 Dec 01 19:42:05 crc kubenswrapper[4960]: I1201 19:42:05.629381 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"face8246-9a4b-4267-b919-2f78a451ba18","Type":"ContainerDied","Data":"1cb851ef1d549cd8ed68433c2123953be674b69744ebe0476a72e95c811bd749"} Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.030677 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:06 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:06 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:06 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.030755 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.769695 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.771038 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.774917 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.778393 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.795412 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.822139 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.822261 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.924206 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.924339 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.924441 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:06 crc kubenswrapper[4960]: I1201 19:42:06.953850 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.024531 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.030385 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:07 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:07 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:07 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.030451 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.106382 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.135805 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir\") pod \"face8246-9a4b-4267-b919-2f78a451ba18\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.135916 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access\") pod \"face8246-9a4b-4267-b919-2f78a451ba18\" (UID: \"face8246-9a4b-4267-b919-2f78a451ba18\") " Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.136021 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "face8246-9a4b-4267-b919-2f78a451ba18" (UID: "face8246-9a4b-4267-b919-2f78a451ba18"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.136273 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/face8246-9a4b-4267-b919-2f78a451ba18-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.140005 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "face8246-9a4b-4267-b919-2f78a451ba18" (UID: "face8246-9a4b-4267-b919-2f78a451ba18"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.237279 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/face8246-9a4b-4267-b919-2f78a451ba18-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.504612 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 01 19:42:07 crc kubenswrapper[4960]: W1201 19:42:07.539713 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda36deef0_8cfe_4cf4_880c_3ad3e02113e2.slice/crio-c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4 WatchSource:0}: Error finding container c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4: Status 404 returned error can't find the container with id c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4 Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.672465 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a36deef0-8cfe-4cf4-880c-3ad3e02113e2","Type":"ContainerStarted","Data":"c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4"} Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.680896 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"face8246-9a4b-4267-b919-2f78a451ba18","Type":"ContainerDied","Data":"1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1"} Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.680961 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ebd0373034df673662609d09ba8810691c789aba2c60fdf308406f36d22baf1" Dec 01 19:42:07 crc kubenswrapper[4960]: I1201 19:42:07.681050 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 01 19:42:08 crc kubenswrapper[4960]: I1201 19:42:08.030589 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:08 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:08 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:08 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:08 crc kubenswrapper[4960]: I1201 19:42:08.030665 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:09 crc kubenswrapper[4960]: I1201 19:42:09.030361 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:09 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:09 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:09 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:09 crc kubenswrapper[4960]: I1201 19:42:09.030460 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:09 crc kubenswrapper[4960]: I1201 19:42:09.709792 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a36deef0-8cfe-4cf4-880c-3ad3e02113e2","Type":"ContainerStarted","Data":"f46de5945aa9a270daf2882ffe8bc0f2a3eda19a66ca1e7b32ec8c372b77f6d6"} Dec 01 19:42:09 crc kubenswrapper[4960]: I1201 19:42:09.748710 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.74868911 podStartE2EDuration="3.74868911s" podCreationTimestamp="2025-12-01 19:42:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:09.73549217 +0000 UTC m=+165.022983839" watchObservedRunningTime="2025-12-01 19:42:09.74868911 +0000 UTC m=+165.036180779" Dec 01 19:42:09 crc kubenswrapper[4960]: I1201 19:42:09.830908 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wv5p9" Dec 01 19:42:10 crc kubenswrapper[4960]: I1201 19:42:10.031434 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:10 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:10 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:10 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:10 crc kubenswrapper[4960]: I1201 19:42:10.031513 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:10 crc kubenswrapper[4960]: I1201 19:42:10.723755 4960 generic.go:334] "Generic (PLEG): container finished" podID="a36deef0-8cfe-4cf4-880c-3ad3e02113e2" containerID="f46de5945aa9a270daf2882ffe8bc0f2a3eda19a66ca1e7b32ec8c372b77f6d6" exitCode=0 Dec 01 19:42:10 crc kubenswrapper[4960]: I1201 19:42:10.723807 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a36deef0-8cfe-4cf4-880c-3ad3e02113e2","Type":"ContainerDied","Data":"f46de5945aa9a270daf2882ffe8bc0f2a3eda19a66ca1e7b32ec8c372b77f6d6"} Dec 01 19:42:11 crc kubenswrapper[4960]: I1201 19:42:11.030574 4960 patch_prober.go:28] interesting pod/router-default-5444994796-dpfnk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 01 19:42:11 crc kubenswrapper[4960]: [-]has-synced failed: reason withheld Dec 01 19:42:11 crc kubenswrapper[4960]: [+]process-running ok Dec 01 19:42:11 crc kubenswrapper[4960]: healthz check failed Dec 01 19:42:11 crc kubenswrapper[4960]: I1201 19:42:11.030661 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dpfnk" podUID="f084aac8-8a48-42db-b4bb-7773174cbd78" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:42:11 crc kubenswrapper[4960]: I1201 19:42:11.226886 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:42:11 crc kubenswrapper[4960]: I1201 19:42:11.245692 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/30d9bdd8-31e7-460b-aefb-421c6d306fbd-metrics-certs\") pod \"network-metrics-daemon-jsbh7\" (UID: \"30d9bdd8-31e7-460b-aefb-421c6d306fbd\") " pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:42:11 crc kubenswrapper[4960]: I1201 19:42:11.349454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jsbh7" Dec 01 19:42:12 crc kubenswrapper[4960]: I1201 19:42:12.031151 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:42:12 crc kubenswrapper[4960]: I1201 19:42:12.034452 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-dpfnk" Dec 01 19:42:13 crc kubenswrapper[4960]: I1201 19:42:13.870641 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:42:13 crc kubenswrapper[4960]: I1201 19:42:13.870699 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:42:13 crc kubenswrapper[4960]: I1201 19:42:13.977869 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:42:13 crc kubenswrapper[4960]: I1201 19:42:13.982823 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:42:14 crc kubenswrapper[4960]: I1201 19:42:14.733189 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-w95l4" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.125104 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.258359 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir\") pod \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.258475 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access\") pod \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\" (UID: \"a36deef0-8cfe-4cf4-880c-3ad3e02113e2\") " Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.258578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a36deef0-8cfe-4cf4-880c-3ad3e02113e2" (UID: "a36deef0-8cfe-4cf4-880c-3ad3e02113e2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.258772 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.267379 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a36deef0-8cfe-4cf4-880c-3ad3e02113e2" (UID: "a36deef0-8cfe-4cf4-880c-3ad3e02113e2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.277716 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jsbh7"] Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.360726 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a36deef0-8cfe-4cf4-880c-3ad3e02113e2-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.800177 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a36deef0-8cfe-4cf4-880c-3ad3e02113e2","Type":"ContainerDied","Data":"c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4"} Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.800218 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 01 19:42:18 crc kubenswrapper[4960]: I1201 19:42:18.800259 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f342a800c163165446533675e8b2fafc541114871e0aa468920f33072bdcc4" Dec 01 19:42:21 crc kubenswrapper[4960]: I1201 19:42:21.986838 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:42:32 crc kubenswrapper[4960]: I1201 19:42:32.822198 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 01 19:42:34 crc kubenswrapper[4960]: I1201 19:42:34.749779 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-j8xsm" Dec 01 19:42:36 crc kubenswrapper[4960]: W1201 19:42:36.349393 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30d9bdd8_31e7_460b_aefb_421c6d306fbd.slice/crio-e70a739102a5a53eec635a1fd572e518ba59c414f20ad9826bf752411b187595 WatchSource:0}: Error finding container e70a739102a5a53eec635a1fd572e518ba59c414f20ad9826bf752411b187595: Status 404 returned error can't find the container with id e70a739102a5a53eec635a1fd572e518ba59c414f20ad9826bf752411b187595 Dec 01 19:42:36 crc kubenswrapper[4960]: E1201 19:42:36.451020 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 19:42:36 crc kubenswrapper[4960]: E1201 19:42:36.451639 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h4wvs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fwhqw_openshift-marketplace(3c55929c-b18d-4719-944e-54176bdb96bb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:36 crc kubenswrapper[4960]: E1201 19:42:36.453157 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fwhqw" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" Dec 01 19:42:36 crc kubenswrapper[4960]: I1201 19:42:36.906650 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" event={"ID":"30d9bdd8-31e7-460b-aefb-421c6d306fbd","Type":"ContainerStarted","Data":"e70a739102a5a53eec635a1fd572e518ba59c414f20ad9826bf752411b187595"} Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.708261 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fwhqw" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.793370 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.793599 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cm2vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lfdlz_openshift-marketplace(1b0cc447-6d51-453e-9eba-ffb99a8a988f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.794754 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lfdlz" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.809909 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.810105 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdsm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-m4x5q_openshift-marketplace(cdba2587-565b-4ffc-9448-5b757cc58410): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.812412 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-m4x5q" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.821222 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.821419 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j2lqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zlm2j_openshift-marketplace(9f2e47de-2360-4506-9e43-ee29da7ca9e1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:37 crc kubenswrapper[4960]: E1201 19:42:37.823509 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zlm2j" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" Dec 01 19:42:38 crc kubenswrapper[4960]: E1201 19:42:38.963977 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lfdlz" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" Dec 01 19:42:38 crc kubenswrapper[4960]: E1201 19:42:38.964295 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-m4x5q" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" Dec 01 19:42:38 crc kubenswrapper[4960]: E1201 19:42:38.964448 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zlm2j" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.038656 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.038843 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9v6mq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kmczg_openshift-marketplace(54d97c64-b099-43d0-803d-0645a7bf5ff9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.040084 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kmczg" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.068313 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.068468 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4lmv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l7ldv_openshift-marketplace(65475df3-6ce0-44d3-a5ed-861e34334058): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.069694 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l7ldv" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.928718 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerID="d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f" exitCode=0 Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.928911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerDied","Data":"d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f"} Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.936469 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" event={"ID":"30d9bdd8-31e7-460b-aefb-421c6d306fbd","Type":"ContainerStarted","Data":"ae17634038cd5fedc24beec7f9ba97233da6a2701fc42afed52c871d0ab20180"} Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.936578 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jsbh7" event={"ID":"30d9bdd8-31e7-460b-aefb-421c6d306fbd","Type":"ContainerStarted","Data":"7c5dd9c80b7766b050cdf0f9aaf917d3473287c3eea9aedd06fef5ab43b7c3e9"} Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.940965 4960 generic.go:334] "Generic (PLEG): container finished" podID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerID="7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1" exitCode=0 Dec 01 19:42:39 crc kubenswrapper[4960]: I1201 19:42:39.941854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerDied","Data":"7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1"} Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.942923 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kmczg" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" Dec 01 19:42:39 crc kubenswrapper[4960]: E1201 19:42:39.943630 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l7ldv" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" Dec 01 19:42:40 crc kubenswrapper[4960]: I1201 19:42:40.084771 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jsbh7" podStartSLOduration=172.084735886 podStartE2EDuration="2m52.084735886s" podCreationTimestamp="2025-12-01 19:39:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:40.050381943 +0000 UTC m=+195.337873632" watchObservedRunningTime="2025-12-01 19:42:40.084735886 +0000 UTC m=+195.372227795" Dec 01 19:42:40 crc kubenswrapper[4960]: I1201 19:42:40.949713 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerStarted","Data":"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be"} Dec 01 19:42:40 crc kubenswrapper[4960]: I1201 19:42:40.952770 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerStarted","Data":"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e"} Dec 01 19:42:40 crc kubenswrapper[4960]: I1201 19:42:40.969627 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jjmnx" podStartSLOduration=2.605360121 podStartE2EDuration="40.969607308s" podCreationTimestamp="2025-12-01 19:42:00 +0000 UTC" firstStartedPulling="2025-12-01 19:42:02.386416934 +0000 UTC m=+157.673908593" lastFinishedPulling="2025-12-01 19:42:40.750664111 +0000 UTC m=+196.038155780" observedRunningTime="2025-12-01 19:42:40.968847164 +0000 UTC m=+196.256338833" watchObservedRunningTime="2025-12-01 19:42:40.969607308 +0000 UTC m=+196.257098977" Dec 01 19:42:40 crc kubenswrapper[4960]: I1201 19:42:40.986221 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lmk2r" podStartSLOduration=3.873446064 podStartE2EDuration="40.986201981s" podCreationTimestamp="2025-12-01 19:42:00 +0000 UTC" firstStartedPulling="2025-12-01 19:42:03.497054469 +0000 UTC m=+158.784546128" lastFinishedPulling="2025-12-01 19:42:40.609810376 +0000 UTC m=+195.897302045" observedRunningTime="2025-12-01 19:42:40.985994095 +0000 UTC m=+196.273485764" watchObservedRunningTime="2025-12-01 19:42:40.986201981 +0000 UTC m=+196.273693650" Dec 01 19:42:41 crc kubenswrapper[4960]: I1201 19:42:41.930380 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:41 crc kubenswrapper[4960]: I1201 19:42:41.930911 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:42 crc kubenswrapper[4960]: I1201 19:42:42.039812 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:42 crc kubenswrapper[4960]: I1201 19:42:42.040001 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.024451 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jjmnx" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="registry-server" probeResult="failure" output=< Dec 01 19:42:43 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 19:42:43 crc kubenswrapper[4960]: > Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.078381 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lmk2r" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="registry-server" probeResult="failure" output=< Dec 01 19:42:43 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 19:42:43 crc kubenswrapper[4960]: > Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.418596 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 19:42:43 crc kubenswrapper[4960]: E1201 19:42:43.418894 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36deef0-8cfe-4cf4-880c-3ad3e02113e2" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.418910 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36deef0-8cfe-4cf4-880c-3ad3e02113e2" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: E1201 19:42:43.418932 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="face8246-9a4b-4267-b919-2f78a451ba18" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.418937 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="face8246-9a4b-4267-b919-2f78a451ba18" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.419062 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36deef0-8cfe-4cf4-880c-3ad3e02113e2" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.419076 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="face8246-9a4b-4267-b919-2f78a451ba18" containerName="pruner" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.419490 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.422981 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.424021 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.433840 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.555583 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.555812 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.656809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.656896 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.657470 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.859894 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.870895 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:42:43 crc kubenswrapper[4960]: I1201 19:42:43.871005 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:42:44 crc kubenswrapper[4960]: I1201 19:42:44.058142 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:44 crc kubenswrapper[4960]: I1201 19:42:44.453843 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 01 19:42:44 crc kubenswrapper[4960]: I1201 19:42:44.977966 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a43e9118-c23b-4fca-bcee-bcd3dde46728","Type":"ContainerStarted","Data":"5d299dcf847bde9086e928e73fb2f62009c228804055de9e68639296e8679f02"} Dec 01 19:42:44 crc kubenswrapper[4960]: I1201 19:42:44.978455 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a43e9118-c23b-4fca-bcee-bcd3dde46728","Type":"ContainerStarted","Data":"eaadf79dd2f5ef6402221f4fa5c9c2004fa5d6af2365a6f7c4c94fbaabb76da8"} Dec 01 19:42:45 crc kubenswrapper[4960]: I1201 19:42:45.007572 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.00754885 podStartE2EDuration="2.00754885s" podCreationTimestamp="2025-12-01 19:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:45.004243817 +0000 UTC m=+200.291735496" watchObservedRunningTime="2025-12-01 19:42:45.00754885 +0000 UTC m=+200.295040519" Dec 01 19:42:45 crc kubenswrapper[4960]: I1201 19:42:45.984352 4960 generic.go:334] "Generic (PLEG): container finished" podID="a43e9118-c23b-4fca-bcee-bcd3dde46728" containerID="5d299dcf847bde9086e928e73fb2f62009c228804055de9e68639296e8679f02" exitCode=0 Dec 01 19:42:45 crc kubenswrapper[4960]: I1201 19:42:45.984408 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a43e9118-c23b-4fca-bcee-bcd3dde46728","Type":"ContainerDied","Data":"5d299dcf847bde9086e928e73fb2f62009c228804055de9e68639296e8679f02"} Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.218111 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.330681 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir\") pod \"a43e9118-c23b-4fca-bcee-bcd3dde46728\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.330859 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access\") pod \"a43e9118-c23b-4fca-bcee-bcd3dde46728\" (UID: \"a43e9118-c23b-4fca-bcee-bcd3dde46728\") " Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.331171 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a43e9118-c23b-4fca-bcee-bcd3dde46728" (UID: "a43e9118-c23b-4fca-bcee-bcd3dde46728"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.344373 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a43e9118-c23b-4fca-bcee-bcd3dde46728" (UID: "a43e9118-c23b-4fca-bcee-bcd3dde46728"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.432167 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a43e9118-c23b-4fca-bcee-bcd3dde46728-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.432214 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a43e9118-c23b-4fca-bcee-bcd3dde46728-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.998020 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a43e9118-c23b-4fca-bcee-bcd3dde46728","Type":"ContainerDied","Data":"eaadf79dd2f5ef6402221f4fa5c9c2004fa5d6af2365a6f7c4c94fbaabb76da8"} Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.998078 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 01 19:42:47 crc kubenswrapper[4960]: I1201 19:42:47.998087 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaadf79dd2f5ef6402221f4fa5c9c2004fa5d6af2365a6f7c4c94fbaabb76da8" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.622165 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 19:42:49 crc kubenswrapper[4960]: E1201 19:42:49.623299 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43e9118-c23b-4fca-bcee-bcd3dde46728" containerName="pruner" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.623337 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43e9118-c23b-4fca-bcee-bcd3dde46728" containerName="pruner" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.623628 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43e9118-c23b-4fca-bcee-bcd3dde46728" containerName="pruner" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.624807 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.628388 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.628488 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.632453 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.664106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.664502 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.664625 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.766671 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.766811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.766858 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.766832 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.767012 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.789964 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access\") pod \"installer-9-crc\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:49 crc kubenswrapper[4960]: I1201 19:42:49.944195 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:42:50 crc kubenswrapper[4960]: I1201 19:42:50.154046 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 01 19:42:51 crc kubenswrapper[4960]: I1201 19:42:51.021225 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"28fb7f2f-982c-453b-abcd-50059e073ebe","Type":"ContainerStarted","Data":"8b0e611432ffd35e59ffaa7c2d7cd39d3fdbfa36053f5855e85fff72834b5169"} Dec 01 19:42:51 crc kubenswrapper[4960]: I1201 19:42:51.974879 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.023867 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.031990 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerStarted","Data":"bc37ae77178bbd76e5cf1cdfacf6c1d8559607cec588b291cabc38725cce6968"} Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.035184 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"28fb7f2f-982c-453b-abcd-50059e073ebe","Type":"ContainerStarted","Data":"704165a6802d04b372898b4ce3e3ceec7ec999bdf0bab7af8dad4af592a19c41"} Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.095286 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.09526201 podStartE2EDuration="3.09526201s" podCreationTimestamp="2025-12-01 19:42:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:42:52.094998842 +0000 UTC m=+207.382490531" watchObservedRunningTime="2025-12-01 19:42:52.09526201 +0000 UTC m=+207.382753679" Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.100437 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:52 crc kubenswrapper[4960]: I1201 19:42:52.140035 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.047403 4960 generic.go:334] "Generic (PLEG): container finished" podID="65475df3-6ce0-44d3-a5ed-861e34334058" containerID="81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de" exitCode=0 Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.047493 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerDied","Data":"81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de"} Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.050944 4960 generic.go:334] "Generic (PLEG): container finished" podID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerID="a0741f0150627b76b6ec91d9964b6dd04e0895845896c5cad28ea509d7bdb53a" exitCode=0 Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.051035 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerDied","Data":"a0741f0150627b76b6ec91d9964b6dd04e0895845896c5cad28ea509d7bdb53a"} Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.054684 4960 generic.go:334] "Generic (PLEG): container finished" podID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerID="999259a8921729ec0761ae460f414646bd0956547f14e52356c4494dc7803c5c" exitCode=0 Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.054771 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerDied","Data":"999259a8921729ec0761ae460f414646bd0956547f14e52356c4494dc7803c5c"} Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.057515 4960 generic.go:334] "Generic (PLEG): container finished" podID="3c55929c-b18d-4719-944e-54176bdb96bb" containerID="bc37ae77178bbd76e5cf1cdfacf6c1d8559607cec588b291cabc38725cce6968" exitCode=0 Dec 01 19:42:53 crc kubenswrapper[4960]: I1201 19:42:53.058294 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerDied","Data":"bc37ae77178bbd76e5cf1cdfacf6c1d8559607cec588b291cabc38725cce6968"} Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.077007 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerStarted","Data":"d775fcaaba693f9fcaa8fe641447a1c290eac7df75a46727b6868911028329c7"} Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.088988 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerStarted","Data":"cb7c20e90cb7e624e5669255605ed8578f5f4017b73474d64c359990a55b48a7"} Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.096710 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerStarted","Data":"4e3b874415d5b694ba4b4f30c50e98ce0218f36051f1ee8d78441d87719b6da4"} Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.111548 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zlm2j" podStartSLOduration=2.82021237 podStartE2EDuration="54.11152497s" podCreationTimestamp="2025-12-01 19:42:00 +0000 UTC" firstStartedPulling="2025-12-01 19:42:02.394596162 +0000 UTC m=+157.682087831" lastFinishedPulling="2025-12-01 19:42:53.685908772 +0000 UTC m=+208.973400431" observedRunningTime="2025-12-01 19:42:54.107418312 +0000 UTC m=+209.394910001" watchObservedRunningTime="2025-12-01 19:42:54.11152497 +0000 UTC m=+209.399016639" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.356354 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.356681 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lmk2r" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="registry-server" containerID="cri-o://906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e" gracePeriod=2 Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.701390 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.754308 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jmxm\" (UniqueName: \"kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm\") pod \"8dd973b5-6ffd-480b-b5b4-49e72325a038\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.754443 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities\") pod \"8dd973b5-6ffd-480b-b5b4-49e72325a038\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.754543 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content\") pod \"8dd973b5-6ffd-480b-b5b4-49e72325a038\" (UID: \"8dd973b5-6ffd-480b-b5b4-49e72325a038\") " Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.755798 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities" (OuterVolumeSpecName: "utilities") pod "8dd973b5-6ffd-480b-b5b4-49e72325a038" (UID: "8dd973b5-6ffd-480b-b5b4-49e72325a038"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.761295 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm" (OuterVolumeSpecName: "kube-api-access-2jmxm") pod "8dd973b5-6ffd-480b-b5b4-49e72325a038" (UID: "8dd973b5-6ffd-480b-b5b4-49e72325a038"). InnerVolumeSpecName "kube-api-access-2jmxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.813734 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dd973b5-6ffd-480b-b5b4-49e72325a038" (UID: "8dd973b5-6ffd-480b-b5b4-49e72325a038"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.855681 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jmxm\" (UniqueName: \"kubernetes.io/projected/8dd973b5-6ffd-480b-b5b4-49e72325a038-kube-api-access-2jmxm\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.855733 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:54 crc kubenswrapper[4960]: I1201 19:42:54.855743 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dd973b5-6ffd-480b-b5b4-49e72325a038-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.105440 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerStarted","Data":"dae88921fdbdd33d7b3a44da52d7a2f4920905f4b6b2ce6d60a8b6fcce686e41"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.107773 4960 generic.go:334] "Generic (PLEG): container finished" podID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerID="906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e" exitCode=0 Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.107857 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerDied","Data":"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.107854 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmk2r" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.107902 4960 scope.go:117] "RemoveContainer" containerID="906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.107888 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmk2r" event={"ID":"8dd973b5-6ffd-480b-b5b4-49e72325a038","Type":"ContainerDied","Data":"13d0a8feebb80bac2dd4d8c8964b67afcd4006c68d165d5399796e7cc8565cce"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.109711 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerID="cb7c20e90cb7e624e5669255605ed8578f5f4017b73474d64c359990a55b48a7" exitCode=0 Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.109765 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerDied","Data":"cb7c20e90cb7e624e5669255605ed8578f5f4017b73474d64c359990a55b48a7"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.116987 4960 generic.go:334] "Generic (PLEG): container finished" podID="cdba2587-565b-4ffc-9448-5b757cc58410" containerID="4e3b874415d5b694ba4b4f30c50e98ce0218f36051f1ee8d78441d87719b6da4" exitCode=0 Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.117097 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerDied","Data":"4e3b874415d5b694ba4b4f30c50e98ce0218f36051f1ee8d78441d87719b6da4"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.122382 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerStarted","Data":"6046f1ee63a0db3d6160afbed536e144f905bf9a2deb52fcd2fcffda1837a51a"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.128095 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerStarted","Data":"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00"} Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.128798 4960 scope.go:117] "RemoveContainer" containerID="d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.132549 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kmczg" podStartSLOduration=2.637449977 podStartE2EDuration="52.132532751s" podCreationTimestamp="2025-12-01 19:42:03 +0000 UTC" firstStartedPulling="2025-12-01 19:42:04.60982379 +0000 UTC m=+159.897315459" lastFinishedPulling="2025-12-01 19:42:54.104906564 +0000 UTC m=+209.392398233" observedRunningTime="2025-12-01 19:42:55.132313284 +0000 UTC m=+210.419804953" watchObservedRunningTime="2025-12-01 19:42:55.132532751 +0000 UTC m=+210.420024420" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.147350 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.150541 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lmk2r"] Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.154030 4960 scope.go:117] "RemoveContainer" containerID="6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.179025 4960 scope.go:117] "RemoveContainer" containerID="906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e" Dec 01 19:42:55 crc kubenswrapper[4960]: E1201 19:42:55.179755 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e\": container with ID starting with 906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e not found: ID does not exist" containerID="906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.179814 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e"} err="failed to get container status \"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e\": rpc error: code = NotFound desc = could not find container \"906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e\": container with ID starting with 906c3972b4ce5289172a8c2b3d041a151d564c6b634c463c5f26719af599fc4e not found: ID does not exist" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.179875 4960 scope.go:117] "RemoveContainer" containerID="d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f" Dec 01 19:42:55 crc kubenswrapper[4960]: E1201 19:42:55.181017 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f\": container with ID starting with d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f not found: ID does not exist" containerID="d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.181053 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f"} err="failed to get container status \"d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f\": rpc error: code = NotFound desc = could not find container \"d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f\": container with ID starting with d19cb78ef10b1ede19cb316c5f388b4b7c3ad1b0bbf8a78bdc31f79848ff422f not found: ID does not exist" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.181074 4960 scope.go:117] "RemoveContainer" containerID="6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c" Dec 01 19:42:55 crc kubenswrapper[4960]: E1201 19:42:55.181645 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c\": container with ID starting with 6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c not found: ID does not exist" containerID="6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.181709 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c"} err="failed to get container status \"6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c\": rpc error: code = NotFound desc = could not find container \"6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c\": container with ID starting with 6c1f4082df95f41999e6815487953f189ef5f6d90ca37f486d761c56db40600c not found: ID does not exist" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.189636 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l7ldv" podStartSLOduration=3.890923336 podStartE2EDuration="53.189613308s" podCreationTimestamp="2025-12-01 19:42:02 +0000 UTC" firstStartedPulling="2025-12-01 19:42:04.597735413 +0000 UTC m=+159.885227082" lastFinishedPulling="2025-12-01 19:42:53.896425385 +0000 UTC m=+209.183917054" observedRunningTime="2025-12-01 19:42:55.186670276 +0000 UTC m=+210.474161945" watchObservedRunningTime="2025-12-01 19:42:55.189613308 +0000 UTC m=+210.477104987" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.205261 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fwhqw" podStartSLOduration=3.930309034 podStartE2EDuration="52.205238974s" podCreationTimestamp="2025-12-01 19:42:03 +0000 UTC" firstStartedPulling="2025-12-01 19:42:05.629135343 +0000 UTC m=+160.916627012" lastFinishedPulling="2025-12-01 19:42:53.904065283 +0000 UTC m=+209.191556952" observedRunningTime="2025-12-01 19:42:55.20479834 +0000 UTC m=+210.492290009" watchObservedRunningTime="2025-12-01 19:42:55.205238974 +0000 UTC m=+210.492730633" Dec 01 19:42:55 crc kubenswrapper[4960]: I1201 19:42:55.331655 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" path="/var/lib/kubelet/pods/8dd973b5-6ffd-480b-b5b4-49e72325a038/volumes" Dec 01 19:42:56 crc kubenswrapper[4960]: I1201 19:42:56.136934 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerStarted","Data":"3b09407d61c3a128d1e25a04f9dfb5231df71d56684def698078ea44368ce530"} Dec 01 19:42:56 crc kubenswrapper[4960]: I1201 19:42:56.138938 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerStarted","Data":"58f8f294204093a8b43107b18cbb92b30fe40df0032f36b7c2989b7274fec071"} Dec 01 19:42:56 crc kubenswrapper[4960]: I1201 19:42:56.158204 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lfdlz" podStartSLOduration=1.80339356 podStartE2EDuration="55.158182317s" podCreationTimestamp="2025-12-01 19:42:01 +0000 UTC" firstStartedPulling="2025-12-01 19:42:02.390712544 +0000 UTC m=+157.678204213" lastFinishedPulling="2025-12-01 19:42:55.745501301 +0000 UTC m=+211.032992970" observedRunningTime="2025-12-01 19:42:56.156557745 +0000 UTC m=+211.444049415" watchObservedRunningTime="2025-12-01 19:42:56.158182317 +0000 UTC m=+211.445673986" Dec 01 19:42:56 crc kubenswrapper[4960]: I1201 19:42:56.178165 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m4x5q" podStartSLOduration=2.242593055 podStartE2EDuration="52.178147988s" podCreationTimestamp="2025-12-01 19:42:04 +0000 UTC" firstStartedPulling="2025-12-01 19:42:05.619790499 +0000 UTC m=+160.907282168" lastFinishedPulling="2025-12-01 19:42:55.555345432 +0000 UTC m=+210.842837101" observedRunningTime="2025-12-01 19:42:56.175016641 +0000 UTC m=+211.462508310" watchObservedRunningTime="2025-12-01 19:42:56.178147988 +0000 UTC m=+211.465639647" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.092365 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.093094 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.145769 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.226606 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.487442 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.487937 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:01 crc kubenswrapper[4960]: I1201 19:43:01.548859 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:02 crc kubenswrapper[4960]: I1201 19:43:02.210830 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:02 crc kubenswrapper[4960]: I1201 19:43:02.758791 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.029724 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.029795 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.098938 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.231608 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.440506 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.440594 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:03 crc kubenswrapper[4960]: I1201 19:43:03.512892 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.028815 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.030097 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.081765 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.186753 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lfdlz" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="registry-server" containerID="cri-o://3b09407d61c3a128d1e25a04f9dfb5231df71d56684def698078ea44368ce530" gracePeriod=2 Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.232147 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.233186 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.458470 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.458801 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:04 crc kubenswrapper[4960]: I1201 19:43:04.505252 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:05 crc kubenswrapper[4960]: I1201 19:43:05.193599 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerID="3b09407d61c3a128d1e25a04f9dfb5231df71d56684def698078ea44368ce530" exitCode=0 Dec 01 19:43:05 crc kubenswrapper[4960]: I1201 19:43:05.193741 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerDied","Data":"3b09407d61c3a128d1e25a04f9dfb5231df71d56684def698078ea44368ce530"} Dec 01 19:43:05 crc kubenswrapper[4960]: I1201 19:43:05.239653 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:05 crc kubenswrapper[4960]: I1201 19:43:05.356594 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.109738 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.125914 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content\") pod \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.126040 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities\") pod \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.126071 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm2vq\" (UniqueName: \"kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq\") pod \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\" (UID: \"1b0cc447-6d51-453e-9eba-ffb99a8a988f\") " Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.126820 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities" (OuterVolumeSpecName: "utilities") pod "1b0cc447-6d51-453e-9eba-ffb99a8a988f" (UID: "1b0cc447-6d51-453e-9eba-ffb99a8a988f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.134312 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq" (OuterVolumeSpecName: "kube-api-access-cm2vq") pod "1b0cc447-6d51-453e-9eba-ffb99a8a988f" (UID: "1b0cc447-6d51-453e-9eba-ffb99a8a988f"). InnerVolumeSpecName "kube-api-access-cm2vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.205553 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lfdlz" event={"ID":"1b0cc447-6d51-453e-9eba-ffb99a8a988f","Type":"ContainerDied","Data":"237af35b8b1800aa903849d6825632d5db942e39aeb79828176f80856d003af9"} Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.205568 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lfdlz" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.205650 4960 scope.go:117] "RemoveContainer" containerID="3b09407d61c3a128d1e25a04f9dfb5231df71d56684def698078ea44368ce530" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.206493 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kmczg" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="registry-server" containerID="cri-o://dae88921fdbdd33d7b3a44da52d7a2f4920905f4b6b2ce6d60a8b6fcce686e41" gracePeriod=2 Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.227191 4960 scope.go:117] "RemoveContainer" containerID="cb7c20e90cb7e624e5669255605ed8578f5f4017b73474d64c359990a55b48a7" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.229350 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.229418 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm2vq\" (UniqueName: \"kubernetes.io/projected/1b0cc447-6d51-453e-9eba-ffb99a8a988f-kube-api-access-cm2vq\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.255017 4960 scope.go:117] "RemoveContainer" containerID="c5c4d2b87c939f6933da8f4e03b843e19dd2f3f957949bb31491d39929a81401" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.267105 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b0cc447-6d51-453e-9eba-ffb99a8a988f" (UID: "1b0cc447-6d51-453e-9eba-ffb99a8a988f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.330938 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b0cc447-6d51-453e-9eba-ffb99a8a988f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.568927 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:43:06 crc kubenswrapper[4960]: I1201 19:43:06.576013 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lfdlz"] Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.154355 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.213996 4960 generic.go:334] "Generic (PLEG): container finished" podID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerID="dae88921fdbdd33d7b3a44da52d7a2f4920905f4b6b2ce6d60a8b6fcce686e41" exitCode=0 Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.214076 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerDied","Data":"dae88921fdbdd33d7b3a44da52d7a2f4920905f4b6b2ce6d60a8b6fcce686e41"} Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.215714 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m4x5q" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="registry-server" containerID="cri-o://58f8f294204093a8b43107b18cbb92b30fe40df0032f36b7c2989b7274fec071" gracePeriod=2 Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.336560 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" path="/var/lib/kubelet/pods/1b0cc447-6d51-453e-9eba-ffb99a8a988f/volumes" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.444397 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.550599 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v6mq\" (UniqueName: \"kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq\") pod \"54d97c64-b099-43d0-803d-0645a7bf5ff9\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.550701 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content\") pod \"54d97c64-b099-43d0-803d-0645a7bf5ff9\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.550767 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities\") pod \"54d97c64-b099-43d0-803d-0645a7bf5ff9\" (UID: \"54d97c64-b099-43d0-803d-0645a7bf5ff9\") " Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.551962 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities" (OuterVolumeSpecName: "utilities") pod "54d97c64-b099-43d0-803d-0645a7bf5ff9" (UID: "54d97c64-b099-43d0-803d-0645a7bf5ff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.557283 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq" (OuterVolumeSpecName: "kube-api-access-9v6mq") pod "54d97c64-b099-43d0-803d-0645a7bf5ff9" (UID: "54d97c64-b099-43d0-803d-0645a7bf5ff9"). InnerVolumeSpecName "kube-api-access-9v6mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.570587 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54d97c64-b099-43d0-803d-0645a7bf5ff9" (UID: "54d97c64-b099-43d0-803d-0645a7bf5ff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.653523 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v6mq\" (UniqueName: \"kubernetes.io/projected/54d97c64-b099-43d0-803d-0645a7bf5ff9-kube-api-access-9v6mq\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.653569 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:07 crc kubenswrapper[4960]: I1201 19:43:07.653582 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54d97c64-b099-43d0-803d-0645a7bf5ff9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.227602 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kmczg" event={"ID":"54d97c64-b099-43d0-803d-0645a7bf5ff9","Type":"ContainerDied","Data":"70a53a1c4fcb710de2cc018a28c714b3df7fca065ef6a21e29fcfe393db82b77"} Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.227771 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kmczg" Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.227760 4960 scope.go:117] "RemoveContainer" containerID="dae88921fdbdd33d7b3a44da52d7a2f4920905f4b6b2ce6d60a8b6fcce686e41" Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.232641 4960 generic.go:334] "Generic (PLEG): container finished" podID="cdba2587-565b-4ffc-9448-5b757cc58410" containerID="58f8f294204093a8b43107b18cbb92b30fe40df0032f36b7c2989b7274fec071" exitCode=0 Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.232720 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerDied","Data":"58f8f294204093a8b43107b18cbb92b30fe40df0032f36b7c2989b7274fec071"} Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.253860 4960 scope.go:117] "RemoveContainer" containerID="a0741f0150627b76b6ec91d9964b6dd04e0895845896c5cad28ea509d7bdb53a" Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.281872 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.284298 4960 scope.go:117] "RemoveContainer" containerID="59aae7d51edbaed8f392ccc805f9a7f05a9fb9c560510832d35cde707bf0d2c0" Dec 01 19:43:08 crc kubenswrapper[4960]: I1201 19:43:08.288943 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kmczg"] Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.331397 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" path="/var/lib/kubelet/pods/54d97c64-b099-43d0-803d-0645a7bf5ff9/volumes" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.388098 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.481364 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content\") pod \"cdba2587-565b-4ffc-9448-5b757cc58410\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.481429 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdsm5\" (UniqueName: \"kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5\") pod \"cdba2587-565b-4ffc-9448-5b757cc58410\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.481475 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities\") pod \"cdba2587-565b-4ffc-9448-5b757cc58410\" (UID: \"cdba2587-565b-4ffc-9448-5b757cc58410\") " Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.482531 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities" (OuterVolumeSpecName: "utilities") pod "cdba2587-565b-4ffc-9448-5b757cc58410" (UID: "cdba2587-565b-4ffc-9448-5b757cc58410"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.487834 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5" (OuterVolumeSpecName: "kube-api-access-sdsm5") pod "cdba2587-565b-4ffc-9448-5b757cc58410" (UID: "cdba2587-565b-4ffc-9448-5b757cc58410"). InnerVolumeSpecName "kube-api-access-sdsm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.582637 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdsm5\" (UniqueName: \"kubernetes.io/projected/cdba2587-565b-4ffc-9448-5b757cc58410-kube-api-access-sdsm5\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.583198 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.611017 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdba2587-565b-4ffc-9448-5b757cc58410" (UID: "cdba2587-565b-4ffc-9448-5b757cc58410"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:43:09 crc kubenswrapper[4960]: I1201 19:43:09.685192 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdba2587-565b-4ffc-9448-5b757cc58410-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.250485 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m4x5q" event={"ID":"cdba2587-565b-4ffc-9448-5b757cc58410","Type":"ContainerDied","Data":"0dc878703ab0491ab2a17ce1207030a426a9f261f799e371ef987877f6e669c4"} Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.250558 4960 scope.go:117] "RemoveContainer" containerID="58f8f294204093a8b43107b18cbb92b30fe40df0032f36b7c2989b7274fec071" Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.250622 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m4x5q" Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.269655 4960 scope.go:117] "RemoveContainer" containerID="4e3b874415d5b694ba4b4f30c50e98ce0218f36051f1ee8d78441d87719b6da4" Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.281413 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.289167 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m4x5q"] Dec 01 19:43:10 crc kubenswrapper[4960]: I1201 19:43:10.305754 4960 scope.go:117] "RemoveContainer" containerID="2e519b96a259b4b2837e9eca5df3feceb9a05ea5780112f84a53702c6eb4b427" Dec 01 19:43:11 crc kubenswrapper[4960]: I1201 19:43:11.338633 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" path="/var/lib/kubelet/pods/cdba2587-565b-4ffc-9448-5b757cc58410/volumes" Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.413198 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-v8sgq"] Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.871386 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.871503 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.871583 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.872674 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 19:43:13 crc kubenswrapper[4960]: I1201 19:43:13.872786 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545" gracePeriod=600 Dec 01 19:43:15 crc kubenswrapper[4960]: I1201 19:43:15.291366 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545" exitCode=0 Dec 01 19:43:15 crc kubenswrapper[4960]: I1201 19:43:15.291441 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545"} Dec 01 19:43:16 crc kubenswrapper[4960]: I1201 19:43:16.303792 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967"} Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.084605 4960 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086353 4960 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086574 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086587 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086599 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086606 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086613 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086619 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086629 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086635 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086644 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086650 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086660 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086665 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086674 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086679 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086689 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086695 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086703 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086709 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="extract-utilities" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086717 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086723 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086733 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086739 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="extract-content" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.086748 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086753 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086871 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d97c64-b099-43d0-803d-0645a7bf5ff9" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086881 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdba2587-565b-4ffc-9448-5b757cc58410" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086889 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd973b5-6ffd-480b-b5b4-49e72325a038" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.086901 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b0cc447-6d51-453e-9eba-ffb99a8a988f" containerName="registry-server" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.087296 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089369 4960 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089623 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2" gracePeriod=15 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089812 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1" gracePeriod=15 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089862 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d" gracePeriod=15 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089897 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790" gracePeriod=15 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.089934 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365" gracePeriod=15 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.092919 4960 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.093655 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.093702 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.094214 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094237 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.094274 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094291 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.094310 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094326 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.094345 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094361 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.094398 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094413 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094741 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094770 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094861 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094947 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094973 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.094996 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.095260 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.095283 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.148933 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175272 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175385 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175436 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175485 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175527 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175567 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175609 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.175670 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277260 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277322 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277378 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277402 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277426 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277440 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277517 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277530 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277466 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277554 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277724 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277554 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277820 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.277517 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.402645 4960 generic.go:334] "Generic (PLEG): container finished" podID="28fb7f2f-982c-453b-abcd-50059e073ebe" containerID="704165a6802d04b372898b4ce3e3ceec7ec999bdf0bab7af8dad4af592a19c41" exitCode=0 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.402817 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"28fb7f2f-982c-453b-abcd-50059e073ebe","Type":"ContainerDied","Data":"704165a6802d04b372898b4ce3e3ceec7ec999bdf0bab7af8dad4af592a19c41"} Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.404363 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.404788 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.407165 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.409473 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.410561 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1" exitCode=0 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.410594 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d" exitCode=0 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.410609 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790" exitCode=0 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.410623 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365" exitCode=2 Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.410693 4960 scope.go:117] "RemoveContainer" containerID="8ae3318ca610719560b275425374a2c4baf945d8b7a5b5c13db7f4c08fbc309a" Dec 01 19:43:29 crc kubenswrapper[4960]: I1201 19:43:29.439642 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:43:29 crc kubenswrapper[4960]: W1201 19:43:29.479158 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-f8d733e0f58b4ecaaaaa6ec5f85d78e2a4097bb84ac48d7aa051fb7339d179fd WatchSource:0}: Error finding container f8d733e0f58b4ecaaaaa6ec5f85d78e2a4097bb84ac48d7aa051fb7339d179fd: Status 404 returned error can't find the container with id f8d733e0f58b4ecaaaaa6ec5f85d78e2a4097bb84ac48d7aa051fb7339d179fd Dec 01 19:43:29 crc kubenswrapper[4960]: E1201 19:43:29.485481 4960 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d2ee984844628 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,LastTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.294553 4960 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.295482 4960 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.296041 4960 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.296574 4960 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.297295 4960 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.297331 4960 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.297653 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="200ms" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.425923 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.429802 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b"} Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.429887 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f8d733e0f58b4ecaaaaa6ec5f85d78e2a4097bb84ac48d7aa051fb7339d179fd"} Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.431223 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.431611 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.499586 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="400ms" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.705984 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.707752 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.709391 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.797973 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access\") pod \"28fb7f2f-982c-453b-abcd-50059e073ebe\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798056 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir\") pod \"28fb7f2f-982c-453b-abcd-50059e073ebe\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798213 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock\") pod \"28fb7f2f-982c-453b-abcd-50059e073ebe\" (UID: \"28fb7f2f-982c-453b-abcd-50059e073ebe\") " Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798276 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "28fb7f2f-982c-453b-abcd-50059e073ebe" (UID: "28fb7f2f-982c-453b-abcd-50059e073ebe"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798492 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock" (OuterVolumeSpecName: "var-lock") pod "28fb7f2f-982c-453b-abcd-50059e073ebe" (UID: "28fb7f2f-982c-453b-abcd-50059e073ebe"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798671 4960 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.798699 4960 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/28fb7f2f-982c-453b-abcd-50059e073ebe-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.806267 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "28fb7f2f-982c-453b-abcd-50059e073ebe" (UID: "28fb7f2f-982c-453b-abcd-50059e073ebe"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:43:30 crc kubenswrapper[4960]: I1201 19:43:30.900584 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28fb7f2f-982c-453b-abcd-50059e073ebe-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:30 crc kubenswrapper[4960]: E1201 19:43:30.901317 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="800ms" Dec 01 19:43:31 crc kubenswrapper[4960]: E1201 19:43:31.389736 4960 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d2ee984844628 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,LastTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 19:43:31 crc kubenswrapper[4960]: I1201 19:43:31.437896 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 01 19:43:31 crc kubenswrapper[4960]: I1201 19:43:31.437913 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"28fb7f2f-982c-453b-abcd-50059e073ebe","Type":"ContainerDied","Data":"8b0e611432ffd35e59ffaa7c2d7cd39d3fdbfa36053f5855e85fff72834b5169"} Dec 01 19:43:31 crc kubenswrapper[4960]: I1201 19:43:31.437973 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b0e611432ffd35e59ffaa7c2d7cd39d3fdbfa36053f5855e85fff72834b5169" Dec 01 19:43:31 crc kubenswrapper[4960]: I1201 19:43:31.442497 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:31 crc kubenswrapper[4960]: I1201 19:43:31.443157 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:31 crc kubenswrapper[4960]: E1201 19:43:31.702707 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="1.6s" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.076261 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.077609 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.078560 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.078945 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.079449 4960 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118039 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118160 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118167 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118284 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118273 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118376 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118914 4960 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118933 4960 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.118948 4960 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.451764 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.453571 4960 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2" exitCode=0 Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.453662 4960 scope.go:117] "RemoveContainer" containerID="f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.453764 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.480694 4960 scope.go:117] "RemoveContainer" containerID="b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.486738 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.487576 4960 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.488383 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.497041 4960 scope.go:117] "RemoveContainer" containerID="95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.514929 4960 scope.go:117] "RemoveContainer" containerID="900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.536617 4960 scope.go:117] "RemoveContainer" containerID="9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.558440 4960 scope.go:117] "RemoveContainer" containerID="cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.593300 4960 scope.go:117] "RemoveContainer" containerID="f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.594096 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\": container with ID starting with f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1 not found: ID does not exist" containerID="f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.594150 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1"} err="failed to get container status \"f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\": rpc error: code = NotFound desc = could not find container \"f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1\": container with ID starting with f8a226870db270320bbcd038cb5a1525f9b02e0a7d617ac5cdc23a81ff2dabd1 not found: ID does not exist" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.594182 4960 scope.go:117] "RemoveContainer" containerID="b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.595136 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\": container with ID starting with b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d not found: ID does not exist" containerID="b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.595186 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d"} err="failed to get container status \"b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\": rpc error: code = NotFound desc = could not find container \"b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d\": container with ID starting with b96339582889250b85d35f911ee931c98db4bd27248d87bfeda5ffece28e310d not found: ID does not exist" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.595219 4960 scope.go:117] "RemoveContainer" containerID="95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.595765 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\": container with ID starting with 95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790 not found: ID does not exist" containerID="95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.595806 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790"} err="failed to get container status \"95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\": rpc error: code = NotFound desc = could not find container \"95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790\": container with ID starting with 95218719dc7363b6ba5f802c38e526b7bb55f1a8b87cffcf27c823f20eaff790 not found: ID does not exist" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.595822 4960 scope.go:117] "RemoveContainer" containerID="900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.596835 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\": container with ID starting with 900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365 not found: ID does not exist" containerID="900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.596877 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365"} err="failed to get container status \"900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\": rpc error: code = NotFound desc = could not find container \"900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365\": container with ID starting with 900a8701f73f55724a970c0d38dd810521fa3dee0f3071258cc469de0b604365 not found: ID does not exist" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.596897 4960 scope.go:117] "RemoveContainer" containerID="9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.597725 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\": container with ID starting with 9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2 not found: ID does not exist" containerID="9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.597755 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2"} err="failed to get container status \"9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\": rpc error: code = NotFound desc = could not find container \"9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2\": container with ID starting with 9731a8ed780506f7d87fd211ae354b890d7ce5eae38cc18f0c4dc1c6c4865ad2 not found: ID does not exist" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.597771 4960 scope.go:117] "RemoveContainer" containerID="cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b" Dec 01 19:43:32 crc kubenswrapper[4960]: E1201 19:43:32.598138 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\": container with ID starting with cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b not found: ID does not exist" containerID="cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b" Dec 01 19:43:32 crc kubenswrapper[4960]: I1201 19:43:32.598169 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b"} err="failed to get container status \"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\": rpc error: code = NotFound desc = could not find container \"cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b\": container with ID starting with cf7fab833d8bf4498c88494cdfac3f1af5bece93890cf0351838887d25784d1b not found: ID does not exist" Dec 01 19:43:33 crc kubenswrapper[4960]: E1201 19:43:33.304731 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="3.2s" Dec 01 19:43:33 crc kubenswrapper[4960]: I1201 19:43:33.339217 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 01 19:43:35 crc kubenswrapper[4960]: I1201 19:43:35.331411 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:35 crc kubenswrapper[4960]: I1201 19:43:35.332045 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:36 crc kubenswrapper[4960]: E1201 19:43:36.507335 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="6.4s" Dec 01 19:43:38 crc kubenswrapper[4960]: E1201 19:43:38.397500 4960 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" volumeName="registry-storage" Dec 01 19:43:38 crc kubenswrapper[4960]: I1201 19:43:38.455555 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" containerName="oauth-openshift" containerID="cri-o://a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c" gracePeriod=15 Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.373663 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.374636 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.375160 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.375569 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427099 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427220 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427337 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427392 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427453 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427488 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427549 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427585 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427626 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427670 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427713 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427750 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.427850 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8j9s\" (UniqueName: \"kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s\") pod \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\" (UID: \"cc354d2f-b7ad-4878-8a1d-ffae0d774db3\") " Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.428454 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.428468 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.428511 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.429523 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.430213 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.434479 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.434761 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.435936 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.436200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s" (OuterVolumeSpecName: "kube-api-access-h8j9s") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "kube-api-access-h8j9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.436276 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.436469 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.437831 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.437857 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.438692 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "cc354d2f-b7ad-4878-8a1d-ffae0d774db3" (UID: "cc354d2f-b7ad-4878-8a1d-ffae0d774db3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.507233 4960 generic.go:334] "Generic (PLEG): container finished" podID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" containerID="a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c" exitCode=0 Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.507327 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.507358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" event={"ID":"cc354d2f-b7ad-4878-8a1d-ffae0d774db3","Type":"ContainerDied","Data":"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c"} Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.507854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" event={"ID":"cc354d2f-b7ad-4878-8a1d-ffae0d774db3","Type":"ContainerDied","Data":"97ac7a3e8707ea5f3c4971ac5e79f040c439749bd1b8f582e8525027b83c0ab4"} Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.507907 4960 scope.go:117] "RemoveContainer" containerID="a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.508611 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.509173 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.509565 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.529677 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.529925 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.529972 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.529993 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530016 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530036 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530054 4960 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530073 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530091 4960 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530213 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530242 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530261 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530279 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530285 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530300 4960 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530489 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8j9s\" (UniqueName: \"kubernetes.io/projected/cc354d2f-b7ad-4878-8a1d-ffae0d774db3-kube-api-access-h8j9s\") on node \"crc\" DevicePath \"\"" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.530710 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.537713 4960 scope.go:117] "RemoveContainer" containerID="a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c" Dec 01 19:43:39 crc kubenswrapper[4960]: E1201 19:43:39.538536 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c\": container with ID starting with a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c not found: ID does not exist" containerID="a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c" Dec 01 19:43:39 crc kubenswrapper[4960]: I1201 19:43:39.538612 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c"} err="failed to get container status \"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c\": rpc error: code = NotFound desc = could not find container \"a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c\": container with ID starting with a46579ff2f06dc2535103098ce5c59865e2f76aa41bb1ada327c47a63f357f5c not found: ID does not exist" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.324329 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.325677 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.326248 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.326777 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.340203 4960 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.340272 4960 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:41 crc kubenswrapper[4960]: E1201 19:43:41.341109 4960 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.341736 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:41 crc kubenswrapper[4960]: E1201 19:43:41.392312 4960 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.177:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d2ee984844628 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,LastTimestamp:2025-12-01 19:43:29.484621352 +0000 UTC m=+244.772113061,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 01 19:43:41 crc kubenswrapper[4960]: I1201 19:43:41.522604 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6754b392e494abe36707dd1ea85922dc14588c4610af14d91c2fe63925edc0c3"} Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.534601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.535089 4960 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57" exitCode=1 Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.535195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57"} Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.535982 4960 scope.go:117] "RemoveContainer" containerID="b12082b600a77fd3168c428364aa535d8800e4e32943916ee5bc19abfd0b8d57" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.536060 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.536321 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.536746 4960 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.537442 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.538093 4960 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="5c7c74c8502dd8372343554424aba0da54918c7ae7ec1983370a2922966799d5" exitCode=0 Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.538139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"5c7c74c8502dd8372343554424aba0da54918c7ae7ec1983370a2922966799d5"} Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.538353 4960 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.538369 4960 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:42 crc kubenswrapper[4960]: E1201 19:43:42.538743 4960 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.538944 4960 status_manager.go:851] "Failed to get status for pod" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.539461 4960 status_manager.go:851] "Failed to get status for pod" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" pod="openshift-authentication/oauth-openshift-558db77b4-v8sgq" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-v8sgq\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.539922 4960 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: I1201 19:43:42.541285 4960 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.177:6443: connect: connection refused" Dec 01 19:43:42 crc kubenswrapper[4960]: E1201 19:43:42.909361 4960 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.177:6443: connect: connection refused" interval="7s" Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.548759 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ad7707dfdadf603af7a327519e07be6dee264f66d30b8161c0b1eb61ab90677f"} Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.549212 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5e9bc31bc6fc44fc3b6ece24679680c0f80d1f8c80509918b4d201fcde298390"} Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.549233 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ad3736e34b2a21b2a31ee477091a5d7b0f69e93fbfe55668b8173f8307642ae1"} Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.549247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"06e7134b974c937f8a24d1bb472eb5148399a0ededa76a257c247b54e8bd67c7"} Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.552646 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 01 19:43:43 crc kubenswrapper[4960]: I1201 19:43:43.552707 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"06bc3f862aa251e46473b80ce5632f49fba99a4f2d9c1434e898342fbda7ec2f"} Dec 01 19:43:44 crc kubenswrapper[4960]: I1201 19:43:44.562405 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"41081f5654b1575fef0667dbd39f000eff1bb440a1926e2f374041071968305b"} Dec 01 19:43:44 crc kubenswrapper[4960]: I1201 19:43:44.562685 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:44 crc kubenswrapper[4960]: I1201 19:43:44.562918 4960 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:44 crc kubenswrapper[4960]: I1201 19:43:44.562959 4960 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:46 crc kubenswrapper[4960]: I1201 19:43:46.342790 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:46 crc kubenswrapper[4960]: I1201 19:43:46.343453 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:46 crc kubenswrapper[4960]: I1201 19:43:46.348546 4960 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]log ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]etcd ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/generic-apiserver-start-informers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-filter ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-apiextensions-informers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-apiextensions-controllers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/crd-informer-synced ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-system-namespaces-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/bootstrap-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/start-kube-aggregator-informers ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-registration-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-discovery-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]autoregister-completion ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-openapi-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 01 19:43:46 crc kubenswrapper[4960]: livez check failed Dec 01 19:43:46 crc kubenswrapper[4960]: I1201 19:43:46.348640 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 01 19:43:49 crc kubenswrapper[4960]: I1201 19:43:49.746986 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:43:49 crc kubenswrapper[4960]: I1201 19:43:49.752573 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.097262 4960 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.233598 4960 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="563eef1d-851f-4af2-a949-0eccf4d268b5" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.308840 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.601582 4960 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.601623 4960 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a4015e4b-ec52-4bf4-a62e-0cc4b6954dba" Dec 01 19:43:50 crc kubenswrapper[4960]: I1201 19:43:50.605248 4960 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="563eef1d-851f-4af2-a949-0eccf4d268b5" Dec 01 19:44:00 crc kubenswrapper[4960]: I1201 19:44:00.315089 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 01 19:44:00 crc kubenswrapper[4960]: I1201 19:44:00.457498 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 01 19:44:00 crc kubenswrapper[4960]: I1201 19:44:00.543073 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 01 19:44:00 crc kubenswrapper[4960]: I1201 19:44:00.734350 4960 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.262499 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.272945 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.382057 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.425444 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.564364 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.644295 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.689194 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 01 19:44:01 crc kubenswrapper[4960]: I1201 19:44:01.696599 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.087043 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.209900 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.288479 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.397502 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.512329 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.603170 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.658239 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.867500 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 01 19:44:02 crc kubenswrapper[4960]: I1201 19:44:02.962249 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.055194 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.133247 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.522419 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.582885 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.593386 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.648665 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.676108 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.736819 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.901087 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 01 19:44:03 crc kubenswrapper[4960]: I1201 19:44:03.935074 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.023877 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.070473 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.076831 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.165035 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.172919 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.189193 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.248221 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.260063 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.261727 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.379536 4960 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.384595 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=35.38456361 podStartE2EDuration="35.38456361s" podCreationTimestamp="2025-12-01 19:43:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:43:50.149088598 +0000 UTC m=+265.436580267" watchObservedRunningTime="2025-12-01 19:44:04.38456361 +0000 UTC m=+279.672055309" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.387503 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-v8sgq"] Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.387613 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.389030 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.392801 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.408344 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.408326659 podStartE2EDuration="14.408326659s" podCreationTimestamp="2025-12-01 19:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:44:04.407632276 +0000 UTC m=+279.695123945" watchObservedRunningTime="2025-12-01 19:44:04.408326659 +0000 UTC m=+279.695818318" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.426212 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.430521 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.835602 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.876806 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.919363 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.931461 4960 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 01 19:44:04 crc kubenswrapper[4960]: I1201 19:44:04.949964 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.026483 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.029444 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.173205 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.187700 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.282862 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.299215 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.331495 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" path="/var/lib/kubelet/pods/cc354d2f-b7ad-4878-8a1d-ffae0d774db3/volumes" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.333874 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.417601 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.435660 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.470405 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.470565 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.525767 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.591854 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.654146 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.667728 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.726532 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.781238 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.836050 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.851177 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.928775 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 01 19:44:05 crc kubenswrapper[4960]: I1201 19:44:05.954285 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.015555 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.058413 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.058849 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.099050 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.153835 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.261176 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.349728 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.355658 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.446366 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.448090 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.503084 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.503680 4960 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.536968 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.542708 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.559110 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.584754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.613011 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.723499 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.817343 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.817343 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.842393 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 01 19:44:06 crc kubenswrapper[4960]: I1201 19:44:06.971863 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.078592 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.084148 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.119509 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.131921 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.249845 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.328873 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.399274 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.453342 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-bjznq"] Dec 01 19:44:07 crc kubenswrapper[4960]: E1201 19:44:07.453791 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" containerName="oauth-openshift" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.453835 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" containerName="oauth-openshift" Dec 01 19:44:07 crc kubenswrapper[4960]: E1201 19:44:07.453892 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" containerName="installer" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.453912 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" containerName="installer" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.454243 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc354d2f-b7ad-4878-8a1d-ffae0d774db3" containerName="oauth-openshift" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.454296 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="28fb7f2f-982c-453b-abcd-50059e073ebe" containerName="installer" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.455159 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.458556 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.461844 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.462152 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.473970 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.474071 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.475013 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.475777 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.475937 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.475952 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.475944 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.476854 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.477511 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.477549 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.487812 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.491555 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.502195 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.519216 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.631614 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.631748 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rhtb\" (UniqueName: \"kubernetes.io/projected/98bcdd9a-9612-4929-ae66-c326a359114e-kube-api-access-5rhtb\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.631910 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632069 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632218 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632418 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632686 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632792 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/98bcdd9a-9612-4929-ae66-c326a359114e-audit-dir\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632882 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.632956 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.633149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-audit-policies\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.633336 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.633579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.633715 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.735989 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736084 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736183 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736234 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/98bcdd9a-9612-4929-ae66-c326a359114e-audit-dir\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736321 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736354 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-audit-policies\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736395 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736444 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736474 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736519 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736553 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rhtb\" (UniqueName: \"kubernetes.io/projected/98bcdd9a-9612-4929-ae66-c326a359114e-kube-api-access-5rhtb\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736586 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.736619 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.738161 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-audit-policies\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.738494 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/98bcdd9a-9612-4929-ae66-c326a359114e-audit-dir\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.738686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.738880 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.740080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.742255 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.748048 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.748256 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.748438 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.748485 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.749085 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.749338 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.751178 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.752251 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/98bcdd9a-9612-4929-ae66-c326a359114e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.767605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rhtb\" (UniqueName: \"kubernetes.io/projected/98bcdd9a-9612-4929-ae66-c326a359114e-kube-api-access-5rhtb\") pod \"oauth-openshift-785f95f6b7-bjznq\" (UID: \"98bcdd9a-9612-4929-ae66-c326a359114e\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.769592 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.788111 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.881452 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.891392 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.892085 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.967437 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 01 19:44:07 crc kubenswrapper[4960]: I1201 19:44:07.973307 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.000175 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.013308 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.014040 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.048109 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.168429 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.192270 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.385927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.404488 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.495474 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.523227 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.574781 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.665000 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.719967 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.764109 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.764761 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.781073 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.842790 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.843473 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.857502 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.858519 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.863393 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 01 19:44:08 crc kubenswrapper[4960]: I1201 19:44:08.902670 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.043771 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.102662 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.146174 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.173484 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.175752 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.193717 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.322724 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.331040 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.382254 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.533226 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.619936 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.672369 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.808053 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 19:44:09 crc kubenswrapper[4960]: I1201 19:44:09.941312 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.037176 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.080880 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.245186 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.255191 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.270488 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.278320 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.288007 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.337650 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.441322 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.475744 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.643530 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.702031 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.704274 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.715439 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.763744 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.801018 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.831593 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.848182 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.968840 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.976536 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 01 19:44:10 crc kubenswrapper[4960]: I1201 19:44:10.976593 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.047819 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.071371 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.116183 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.134175 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.158514 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.189918 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.273667 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.307538 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.315429 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.390063 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.478372 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.569177 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.721969 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.741054 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.873898 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.880675 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 19:44:11 crc kubenswrapper[4960]: I1201 19:44:11.919446 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.019453 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.070005 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.122288 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.181356 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.193305 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.211097 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.271101 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.331340 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.359312 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.415074 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.460951 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.467294 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.549520 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.594010 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.716471 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.731296 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.825366 4960 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.825704 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b" gracePeriod=5 Dec 01 19:44:12 crc kubenswrapper[4960]: I1201 19:44:12.873031 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.097533 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.103888 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.131862 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.202387 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.238961 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-bjznq"] Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.379337 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.612753 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.650097 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: E1201 19:44:13.747338 4960 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 19:44:13 crc kubenswrapper[4960]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-785f95f6b7-bjznq_openshift-authentication_98bcdd9a-9612-4929-ae66-c326a359114e_0(053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e): error adding pod openshift-authentication_oauth-openshift-785f95f6b7-bjznq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e" Netns:"/var/run/netns/8ef3b423-e896-4217-8104-2c0acdec7e65" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-785f95f6b7-bjznq;K8S_POD_INFRA_CONTAINER_ID=053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e;K8S_POD_UID=98bcdd9a-9612-4929-ae66-c326a359114e" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-785f95f6b7-bjznq] networking: Multus: [openshift-authentication/oauth-openshift-785f95f6b7-bjznq/98bcdd9a-9612-4929-ae66-c326a359114e]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-785f95f6b7-bjznq in out of cluster comm: pod "oauth-openshift-785f95f6b7-bjznq" not found Dec 01 19:44:13 crc kubenswrapper[4960]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 19:44:13 crc kubenswrapper[4960]: > Dec 01 19:44:13 crc kubenswrapper[4960]: E1201 19:44:13.747441 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 19:44:13 crc kubenswrapper[4960]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-785f95f6b7-bjznq_openshift-authentication_98bcdd9a-9612-4929-ae66-c326a359114e_0(053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e): error adding pod openshift-authentication_oauth-openshift-785f95f6b7-bjznq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e" Netns:"/var/run/netns/8ef3b423-e896-4217-8104-2c0acdec7e65" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-785f95f6b7-bjznq;K8S_POD_INFRA_CONTAINER_ID=053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e;K8S_POD_UID=98bcdd9a-9612-4929-ae66-c326a359114e" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-785f95f6b7-bjznq] networking: Multus: [openshift-authentication/oauth-openshift-785f95f6b7-bjznq/98bcdd9a-9612-4929-ae66-c326a359114e]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-785f95f6b7-bjznq in out of cluster comm: pod "oauth-openshift-785f95f6b7-bjznq" not found Dec 01 19:44:13 crc kubenswrapper[4960]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 19:44:13 crc kubenswrapper[4960]: > pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:13 crc kubenswrapper[4960]: E1201 19:44:13.747470 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 01 19:44:13 crc kubenswrapper[4960]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-785f95f6b7-bjznq_openshift-authentication_98bcdd9a-9612-4929-ae66-c326a359114e_0(053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e): error adding pod openshift-authentication_oauth-openshift-785f95f6b7-bjznq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e" Netns:"/var/run/netns/8ef3b423-e896-4217-8104-2c0acdec7e65" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-785f95f6b7-bjznq;K8S_POD_INFRA_CONTAINER_ID=053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e;K8S_POD_UID=98bcdd9a-9612-4929-ae66-c326a359114e" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-785f95f6b7-bjznq] networking: Multus: [openshift-authentication/oauth-openshift-785f95f6b7-bjznq/98bcdd9a-9612-4929-ae66-c326a359114e]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-785f95f6b7-bjznq in out of cluster comm: pod "oauth-openshift-785f95f6b7-bjznq" not found Dec 01 19:44:13 crc kubenswrapper[4960]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 19:44:13 crc kubenswrapper[4960]: > pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:13 crc kubenswrapper[4960]: E1201 19:44:13.747547 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-785f95f6b7-bjznq_openshift-authentication(98bcdd9a-9612-4929-ae66-c326a359114e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-785f95f6b7-bjznq_openshift-authentication(98bcdd9a-9612-4929-ae66-c326a359114e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-785f95f6b7-bjznq_openshift-authentication_98bcdd9a-9612-4929-ae66-c326a359114e_0(053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e): error adding pod openshift-authentication_oauth-openshift-785f95f6b7-bjznq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e\\\" Netns:\\\"/var/run/netns/8ef3b423-e896-4217-8104-2c0acdec7e65\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-785f95f6b7-bjznq;K8S_POD_INFRA_CONTAINER_ID=053a6e9b07cc2deb1457248d7bcb01c0e6040a4f90cc89a7ad6914f786f9da6e;K8S_POD_UID=98bcdd9a-9612-4929-ae66-c326a359114e\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-785f95f6b7-bjznq] networking: Multus: [openshift-authentication/oauth-openshift-785f95f6b7-bjznq/98bcdd9a-9612-4929-ae66-c326a359114e]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-785f95f6b7-bjznq in out of cluster comm: pod \\\"oauth-openshift-785f95f6b7-bjznq\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" podUID="98bcdd9a-9612-4929-ae66-c326a359114e" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.757318 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.757805 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.889814 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.895808 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 01 19:44:13 crc kubenswrapper[4960]: I1201 19:44:13.941331 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.017657 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.097530 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.218445 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.311524 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.389465 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.434840 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.509226 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.514987 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.576099 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.613734 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.623994 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.888584 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.935897 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.936369 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 01 19:44:14 crc kubenswrapper[4960]: I1201 19:44:14.942175 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.137521 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.163423 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.172164 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.363887 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.403255 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.487030 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.536456 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.541635 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.695568 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.740199 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 01 19:44:15 crc kubenswrapper[4960]: I1201 19:44:15.875769 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.142581 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.204612 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-bjznq"] Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.440082 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.762212 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.775135 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" event={"ID":"98bcdd9a-9612-4929-ae66-c326a359114e","Type":"ContainerStarted","Data":"f5ebe3f8d5ba2d91f3c706391149ae0c5ce8ef2e5b2b24548ad7aec9d12ba7b6"} Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.775258 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" event={"ID":"98bcdd9a-9612-4929-ae66-c326a359114e","Type":"ContainerStarted","Data":"ae2e1092e74c8dabbb1191cee076e320da15a4c32e7b88c84dd14b2538340935"} Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.776733 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.784152 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" Dec 01 19:44:16 crc kubenswrapper[4960]: I1201 19:44:16.794806 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-785f95f6b7-bjznq" podStartSLOduration=63.794788799 podStartE2EDuration="1m3.794788799s" podCreationTimestamp="2025-12-01 19:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:44:16.793169156 +0000 UTC m=+292.080660835" watchObservedRunningTime="2025-12-01 19:44:16.794788799 +0000 UTC m=+292.082280468" Dec 01 19:44:17 crc kubenswrapper[4960]: I1201 19:44:17.557136 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.043930 4960 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.428053 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.428155 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597500 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597678 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597701 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597715 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597736 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597768 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597774 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.597823 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.598320 4960 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.598357 4960 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.598385 4960 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.598403 4960 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.610746 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.699693 4960 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.793890 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.793956 4960 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b" exitCode=137 Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.794314 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.794476 4960 scope.go:117] "RemoveContainer" containerID="e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.824789 4960 scope.go:117] "RemoveContainer" containerID="e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b" Dec 01 19:44:18 crc kubenswrapper[4960]: E1201 19:44:18.825256 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b\": container with ID starting with e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b not found: ID does not exist" containerID="e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b" Dec 01 19:44:18 crc kubenswrapper[4960]: I1201 19:44:18.825307 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b"} err="failed to get container status \"e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b\": rpc error: code = NotFound desc = could not find container \"e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b\": container with ID starting with e13413e0c05d128c2f9babb2ab25c14b49a0f13755ac4d1523a6cdff31d65e2b not found: ID does not exist" Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.343013 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.344143 4960 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.356013 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.356142 4960 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="647a0deb-2b83-489e-8846-fa2b5353ee18" Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.359428 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 01 19:44:19 crc kubenswrapper[4960]: I1201 19:44:19.359491 4960 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="647a0deb-2b83-489e-8846-fa2b5353ee18" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.088539 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.089828 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" podUID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" containerName="controller-manager" containerID="cri-o://8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11" gracePeriod=30 Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.215314 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.216156 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" podUID="156be4ee-5caf-4423-82aa-9607f9abc19d" containerName="route-controller-manager" containerID="cri-o://825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12" gracePeriod=30 Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.446518 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.473238 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles\") pod \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.473459 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config\") pod \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.473491 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca\") pod \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.473513 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn4j7\" (UniqueName: \"kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7\") pod \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.473553 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert\") pod \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\" (UID: \"9750b5a8-c4e9-45be-aac3-8845bff7bc9b\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.474579 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9750b5a8-c4e9-45be-aac3-8845bff7bc9b" (UID: "9750b5a8-c4e9-45be-aac3-8845bff7bc9b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.474613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca" (OuterVolumeSpecName: "client-ca") pod "9750b5a8-c4e9-45be-aac3-8845bff7bc9b" (UID: "9750b5a8-c4e9-45be-aac3-8845bff7bc9b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.476742 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config" (OuterVolumeSpecName: "config") pod "9750b5a8-c4e9-45be-aac3-8845bff7bc9b" (UID: "9750b5a8-c4e9-45be-aac3-8845bff7bc9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.483339 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7" (OuterVolumeSpecName: "kube-api-access-tn4j7") pod "9750b5a8-c4e9-45be-aac3-8845bff7bc9b" (UID: "9750b5a8-c4e9-45be-aac3-8845bff7bc9b"). InnerVolumeSpecName "kube-api-access-tn4j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.487007 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9750b5a8-c4e9-45be-aac3-8845bff7bc9b" (UID: "9750b5a8-c4e9-45be-aac3-8845bff7bc9b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.538735 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.574727 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca\") pod \"156be4ee-5caf-4423-82aa-9607f9abc19d\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.574842 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert\") pod \"156be4ee-5caf-4423-82aa-9607f9abc19d\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.574917 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config\") pod \"156be4ee-5caf-4423-82aa-9607f9abc19d\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.574979 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2xls\" (UniqueName: \"kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls\") pod \"156be4ee-5caf-4423-82aa-9607f9abc19d\" (UID: \"156be4ee-5caf-4423-82aa-9607f9abc19d\") " Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575309 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575320 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575332 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn4j7\" (UniqueName: \"kubernetes.io/projected/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-kube-api-access-tn4j7\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575343 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575351 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9750b5a8-c4e9-45be-aac3-8845bff7bc9b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.575856 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca" (OuterVolumeSpecName: "client-ca") pod "156be4ee-5caf-4423-82aa-9607f9abc19d" (UID: "156be4ee-5caf-4423-82aa-9607f9abc19d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.576039 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config" (OuterVolumeSpecName: "config") pod "156be4ee-5caf-4423-82aa-9607f9abc19d" (UID: "156be4ee-5caf-4423-82aa-9607f9abc19d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.580492 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls" (OuterVolumeSpecName: "kube-api-access-v2xls") pod "156be4ee-5caf-4423-82aa-9607f9abc19d" (UID: "156be4ee-5caf-4423-82aa-9607f9abc19d"). InnerVolumeSpecName "kube-api-access-v2xls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.580962 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "156be4ee-5caf-4423-82aa-9607f9abc19d" (UID: "156be4ee-5caf-4423-82aa-9607f9abc19d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.677279 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.677322 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2xls\" (UniqueName: \"kubernetes.io/projected/156be4ee-5caf-4423-82aa-9607f9abc19d-kube-api-access-v2xls\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.677333 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/156be4ee-5caf-4423-82aa-9607f9abc19d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.677343 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/156be4ee-5caf-4423-82aa-9607f9abc19d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.898669 4960 generic.go:334] "Generic (PLEG): container finished" podID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" containerID="8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11" exitCode=0 Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.898746 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.898773 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" event={"ID":"9750b5a8-c4e9-45be-aac3-8845bff7bc9b","Type":"ContainerDied","Data":"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11"} Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.899377 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q28sf" event={"ID":"9750b5a8-c4e9-45be-aac3-8845bff7bc9b","Type":"ContainerDied","Data":"4b1f61b69b77140ff3cfa22b7959651903eeee802e31d630293cbf4d9100db86"} Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.899399 4960 scope.go:117] "RemoveContainer" containerID="8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.905324 4960 generic.go:334] "Generic (PLEG): container finished" podID="156be4ee-5caf-4423-82aa-9607f9abc19d" containerID="825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12" exitCode=0 Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.905393 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" event={"ID":"156be4ee-5caf-4423-82aa-9607f9abc19d","Type":"ContainerDied","Data":"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12"} Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.905435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" event={"ID":"156be4ee-5caf-4423-82aa-9607f9abc19d","Type":"ContainerDied","Data":"d78576447b25c7ab2ac76890745d2317eeee7f647095dd0e3803252c40d02479"} Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.905536 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.925868 4960 scope.go:117] "RemoveContainer" containerID="8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11" Dec 01 19:44:30 crc kubenswrapper[4960]: E1201 19:44:30.926375 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11\": container with ID starting with 8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11 not found: ID does not exist" containerID="8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.926421 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11"} err="failed to get container status \"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11\": rpc error: code = NotFound desc = could not find container \"8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11\": container with ID starting with 8c28efca53b43ce0515fc9c52cc6516a57c639ce80f1dc1e050d7efca409cf11 not found: ID does not exist" Dec 01 19:44:30 crc kubenswrapper[4960]: I1201 19:44:30.926454 4960 scope.go:117] "RemoveContainer" containerID="825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.003342 4960 scope.go:117] "RemoveContainer" containerID="825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12" Dec 01 19:44:31 crc kubenswrapper[4960]: E1201 19:44:31.004007 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12\": container with ID starting with 825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12 not found: ID does not exist" containerID="825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.004047 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12"} err="failed to get container status \"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12\": rpc error: code = NotFound desc = could not find container \"825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12\": container with ID starting with 825beaea8196f750c7da9bda8f51e2e84ad6d7a169859f5c7ec10ddc59258d12 not found: ID does not exist" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.006707 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.010133 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dx2ds"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.019855 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.024290 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q28sf"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.335361 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156be4ee-5caf-4423-82aa-9607f9abc19d" path="/var/lib/kubelet/pods/156be4ee-5caf-4423-82aa-9607f9abc19d/volumes" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.336788 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" path="/var/lib/kubelet/pods/9750b5a8-c4e9-45be-aac3-8845bff7bc9b/volumes" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798183 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:44:31 crc kubenswrapper[4960]: E1201 19:44:31.798556 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" containerName="controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798573 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" containerName="controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: E1201 19:44:31.798589 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798597 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 19:44:31 crc kubenswrapper[4960]: E1201 19:44:31.798616 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156be4ee-5caf-4423-82aa-9607f9abc19d" containerName="route-controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798625 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="156be4ee-5caf-4423-82aa-9607f9abc19d" containerName="route-controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798805 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798822 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="156be4ee-5caf-4423-82aa-9607f9abc19d" containerName="route-controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.798836 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9750b5a8-c4e9-45be-aac3-8845bff7bc9b" containerName="controller-manager" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.799370 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.801883 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802071 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802161 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802398 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802471 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802069 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.802702 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.803283 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.805671 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.805727 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.805976 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.806227 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.806440 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.807061 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.817212 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.820678 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.826489 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.895739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.895828 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.895889 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.895937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd2ps\" (UniqueName: \"kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.895975 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.896014 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j75kz\" (UniqueName: \"kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.896057 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.896103 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.896180 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.915723 4960 generic.go:334] "Generic (PLEG): container finished" podID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerID="90d2540886ef78985b81521d66781c786eb822df3aa3ec28b19ea93c432d7624" exitCode=0 Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.915820 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerDied","Data":"90d2540886ef78985b81521d66781c786eb822df3aa3ec28b19ea93c432d7624"} Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.916415 4960 scope.go:117] "RemoveContainer" containerID="90d2540886ef78985b81521d66781c786eb822df3aa3ec28b19ea93c432d7624" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd2ps\" (UniqueName: \"kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997543 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997585 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j75kz\" (UniqueName: \"kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997628 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.997708 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.998912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.998956 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.998964 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:31 crc kubenswrapper[4960]: I1201 19:44:31.999180 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:31.999757 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.001260 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.001462 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.003368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.004634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.004703 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.016485 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j75kz\" (UniqueName: \"kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz\") pod \"route-controller-manager-6ccb45c985-lwqlh\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.018054 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd2ps\" (UniqueName: \"kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps\") pod \"controller-manager-7766fdf5d9-7nrd8\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.129293 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.144403 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.345381 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.413844 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:44:32 crc kubenswrapper[4960]: W1201 19:44:32.426820 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24d71258_296b_4a3c_b54c_ada2e334f9ca.slice/crio-84df0853123fe2373731bb067a8b4eee1997c76a8bc2100ddf1052261862a6ac WatchSource:0}: Error finding container 84df0853123fe2373731bb067a8b4eee1997c76a8bc2100ddf1052261862a6ac: Status 404 returned error can't find the container with id 84df0853123fe2373731bb067a8b4eee1997c76a8bc2100ddf1052261862a6ac Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.923764 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" event={"ID":"d15ad974-a4c4-4a77-9a4a-470f34e53431","Type":"ContainerStarted","Data":"20b095bf7fda0ca7346692005f8329d69d71bf9677e4825beb65e368e1bf802f"} Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.924193 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.924211 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" event={"ID":"d15ad974-a4c4-4a77-9a4a-470f34e53431","Type":"ContainerStarted","Data":"8ad7458c32e9fda7041b57cf1262184e21e458195cf9634f03d8ef739915e0f8"} Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.926664 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerStarted","Data":"0201347d563154d6d9b37e714f6edbf824617b5930fd30d27ebe6a61be27addb"} Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.927477 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.928831 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" event={"ID":"24d71258-296b-4a3c-b54c-ada2e334f9ca","Type":"ContainerStarted","Data":"8474b4a7f671fe29631cb74fea4caa7a1e44f2d956fdafdbe5129c0b8ca10bb1"} Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.928863 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" event={"ID":"24d71258-296b-4a3c-b54c-ada2e334f9ca","Type":"ContainerStarted","Data":"84df0853123fe2373731bb067a8b4eee1997c76a8bc2100ddf1052261862a6ac"} Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.929412 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.929728 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.931384 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:44:32 crc kubenswrapper[4960]: I1201 19:44:32.940414 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" podStartSLOduration=2.94039663 podStartE2EDuration="2.94039663s" podCreationTimestamp="2025-12-01 19:44:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:44:32.938662534 +0000 UTC m=+308.226154213" watchObservedRunningTime="2025-12-01 19:44:32.94039663 +0000 UTC m=+308.227888299" Dec 01 19:44:33 crc kubenswrapper[4960]: I1201 19:44:33.000319 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" podStartSLOduration=3.0002935 podStartE2EDuration="3.0002935s" podCreationTimestamp="2025-12-01 19:44:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:44:32.999432082 +0000 UTC m=+308.286923751" watchObservedRunningTime="2025-12-01 19:44:33.0002935 +0000 UTC m=+308.287785179" Dec 01 19:44:33 crc kubenswrapper[4960]: I1201 19:44:33.005830 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.189016 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8"] Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.190911 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.193615 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.193847 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.198414 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8"] Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.251427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpxdl\" (UniqueName: \"kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.251499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.251645 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.352593 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpxdl\" (UniqueName: \"kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.352690 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.352725 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.353558 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.370686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.387498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpxdl\" (UniqueName: \"kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl\") pod \"collect-profiles-29410305-q7jp8\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:00 crc kubenswrapper[4960]: I1201 19:45:00.513823 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:01 crc kubenswrapper[4960]: I1201 19:45:01.029444 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8"] Dec 01 19:45:01 crc kubenswrapper[4960]: I1201 19:45:01.130968 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" event={"ID":"1beb3eb3-cb74-437d-9c99-faf228d0e172","Type":"ContainerStarted","Data":"096cadcc61e0896d9e4352cdf9cd1d79a685dc8b11e7e8ecdfc1b89539b2fc2b"} Dec 01 19:45:02 crc kubenswrapper[4960]: I1201 19:45:02.139764 4960 generic.go:334] "Generic (PLEG): container finished" podID="1beb3eb3-cb74-437d-9c99-faf228d0e172" containerID="e87ae19577a2513888ea613216aa446c11b0d66c07788f7d54fec47832eb1dc6" exitCode=0 Dec 01 19:45:02 crc kubenswrapper[4960]: I1201 19:45:02.139840 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" event={"ID":"1beb3eb3-cb74-437d-9c99-faf228d0e172","Type":"ContainerDied","Data":"e87ae19577a2513888ea613216aa446c11b0d66c07788f7d54fec47832eb1dc6"} Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.566872 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.709182 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpxdl\" (UniqueName: \"kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl\") pod \"1beb3eb3-cb74-437d-9c99-faf228d0e172\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.709763 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume\") pod \"1beb3eb3-cb74-437d-9c99-faf228d0e172\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.709858 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume\") pod \"1beb3eb3-cb74-437d-9c99-faf228d0e172\" (UID: \"1beb3eb3-cb74-437d-9c99-faf228d0e172\") " Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.711086 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume" (OuterVolumeSpecName: "config-volume") pod "1beb3eb3-cb74-437d-9c99-faf228d0e172" (UID: "1beb3eb3-cb74-437d-9c99-faf228d0e172"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.717156 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl" (OuterVolumeSpecName: "kube-api-access-cpxdl") pod "1beb3eb3-cb74-437d-9c99-faf228d0e172" (UID: "1beb3eb3-cb74-437d-9c99-faf228d0e172"). InnerVolumeSpecName "kube-api-access-cpxdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.727384 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1beb3eb3-cb74-437d-9c99-faf228d0e172" (UID: "1beb3eb3-cb74-437d-9c99-faf228d0e172"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.811278 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpxdl\" (UniqueName: \"kubernetes.io/projected/1beb3eb3-cb74-437d-9c99-faf228d0e172-kube-api-access-cpxdl\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.811815 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1beb3eb3-cb74-437d-9c99-faf228d0e172-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:03 crc kubenswrapper[4960]: I1201 19:45:03.811830 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1beb3eb3-cb74-437d-9c99-faf228d0e172-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.154760 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" event={"ID":"1beb3eb3-cb74-437d-9c99-faf228d0e172","Type":"ContainerDied","Data":"096cadcc61e0896d9e4352cdf9cd1d79a685dc8b11e7e8ecdfc1b89539b2fc2b"} Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.154816 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="096cadcc61e0896d9e4352cdf9cd1d79a685dc8b11e7e8ecdfc1b89539b2fc2b" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.154885 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.572444 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.572809 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jjmnx" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="registry-server" containerID="cri-o://f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be" gracePeriod=30 Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.591957 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.592318 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zlm2j" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="registry-server" containerID="cri-o://d775fcaaba693f9fcaa8fe641447a1c290eac7df75a46727b6868911028329c7" gracePeriod=30 Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.597506 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.597801 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" containerID="cri-o://0201347d563154d6d9b37e714f6edbf824617b5930fd30d27ebe6a61be27addb" gracePeriod=30 Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.601909 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.602237 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l7ldv" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="registry-server" containerID="cri-o://4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00" gracePeriod=30 Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.607955 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:45:04 crc kubenswrapper[4960]: E1201 19:45:04.608192 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1beb3eb3-cb74-437d-9c99-faf228d0e172" containerName="collect-profiles" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.608209 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1beb3eb3-cb74-437d-9c99-faf228d0e172" containerName="collect-profiles" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.608315 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1beb3eb3-cb74-437d-9c99-faf228d0e172" containerName="collect-profiles" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.608682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.623240 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.623322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.623423 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85q84\" (UniqueName: \"kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.642063 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.644039 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fwhqw" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="registry-server" containerID="cri-o://6046f1ee63a0db3d6160afbed536e144f905bf9a2deb52fcd2fcffda1837a51a" gracePeriod=30 Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.656267 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.707841 4960 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w6wcz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.707916 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.724061 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.724162 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.724230 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85q84\" (UniqueName: \"kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.726730 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.737733 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.767895 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85q84\" (UniqueName: \"kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84\") pod \"marketplace-operator-79b997595-wckns\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:04 crc kubenswrapper[4960]: I1201 19:45:04.945072 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.105723 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.107761 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.162790 4960 generic.go:334] "Generic (PLEG): container finished" podID="65475df3-6ce0-44d3-a5ed-861e34334058" containerID="4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00" exitCode=0 Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.162874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerDied","Data":"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.162912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7ldv" event={"ID":"65475df3-6ce0-44d3-a5ed-861e34334058","Type":"ContainerDied","Data":"1ff3fb068b0fb8bc4087ede0409e138a23470af64d25d507c6e49db6ee8bdfd6"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.162936 4960 scope.go:117] "RemoveContainer" containerID="4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.163161 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7ldv" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.166540 4960 generic.go:334] "Generic (PLEG): container finished" podID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerID="f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be" exitCode=0 Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.166660 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjmnx" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.166706 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerDied","Data":"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.166747 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjmnx" event={"ID":"d3901447-b91d-4b95-861d-8e6caf7aa121","Type":"ContainerDied","Data":"819bb0b87e374111614b6b8fbd9ece498dd95f4144025034a419be7af43225cf"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.170091 4960 generic.go:334] "Generic (PLEG): container finished" podID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerID="0201347d563154d6d9b37e714f6edbf824617b5930fd30d27ebe6a61be27addb" exitCode=0 Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.170165 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerDied","Data":"0201347d563154d6d9b37e714f6edbf824617b5930fd30d27ebe6a61be27addb"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.179316 4960 scope.go:117] "RemoveContainer" containerID="81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.179721 4960 generic.go:334] "Generic (PLEG): container finished" podID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerID="d775fcaaba693f9fcaa8fe641447a1c290eac7df75a46727b6868911028329c7" exitCode=0 Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.179815 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerDied","Data":"d775fcaaba693f9fcaa8fe641447a1c290eac7df75a46727b6868911028329c7"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.185176 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerDied","Data":"6046f1ee63a0db3d6160afbed536e144f905bf9a2deb52fcd2fcffda1837a51a"} Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.185463 4960 generic.go:334] "Generic (PLEG): container finished" podID="3c55929c-b18d-4719-944e-54176bdb96bb" containerID="6046f1ee63a0db3d6160afbed536e144f905bf9a2deb52fcd2fcffda1837a51a" exitCode=0 Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.198047 4960 scope.go:117] "RemoveContainer" containerID="cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.210648 4960 scope.go:117] "RemoveContainer" containerID="4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.211416 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00\": container with ID starting with 4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00 not found: ID does not exist" containerID="4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.211499 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00"} err="failed to get container status \"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00\": rpc error: code = NotFound desc = could not find container \"4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00\": container with ID starting with 4d8205831eaf823f2a0f34507f2b6fc8fba52e8e312f084b8500b36321f62d00 not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.211550 4960 scope.go:117] "RemoveContainer" containerID="81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.212014 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de\": container with ID starting with 81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de not found: ID does not exist" containerID="81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.212059 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de"} err="failed to get container status \"81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de\": rpc error: code = NotFound desc = could not find container \"81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de\": container with ID starting with 81c5eaaafe42341ea730996fa60f3f07cb68f73eac2a6e4bf08191148bf6b4de not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.212093 4960 scope.go:117] "RemoveContainer" containerID="cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.212466 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702\": container with ID starting with cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702 not found: ID does not exist" containerID="cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.212516 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702"} err="failed to get container status \"cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702\": rpc error: code = NotFound desc = could not find container \"cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702\": container with ID starting with cb22d9303433e71ba191eaf7e2ba305ae48fc7f111b2b83bab76274c58639702 not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.212573 4960 scope.go:117] "RemoveContainer" containerID="f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.226483 4960 scope.go:117] "RemoveContainer" containerID="7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.228801 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities\") pod \"d3901447-b91d-4b95-861d-8e6caf7aa121\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.228854 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content\") pod \"d3901447-b91d-4b95-861d-8e6caf7aa121\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.228955 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities\") pod \"65475df3-6ce0-44d3-a5ed-861e34334058\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.228989 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content\") pod \"65475df3-6ce0-44d3-a5ed-861e34334058\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.229061 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtblk\" (UniqueName: \"kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk\") pod \"d3901447-b91d-4b95-861d-8e6caf7aa121\" (UID: \"d3901447-b91d-4b95-861d-8e6caf7aa121\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.229083 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lmv2\" (UniqueName: \"kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2\") pod \"65475df3-6ce0-44d3-a5ed-861e34334058\" (UID: \"65475df3-6ce0-44d3-a5ed-861e34334058\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.230251 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities" (OuterVolumeSpecName: "utilities") pod "d3901447-b91d-4b95-861d-8e6caf7aa121" (UID: "d3901447-b91d-4b95-861d-8e6caf7aa121"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.230578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities" (OuterVolumeSpecName: "utilities") pod "65475df3-6ce0-44d3-a5ed-861e34334058" (UID: "65475df3-6ce0-44d3-a5ed-861e34334058"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.235943 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk" (OuterVolumeSpecName: "kube-api-access-gtblk") pod "d3901447-b91d-4b95-861d-8e6caf7aa121" (UID: "d3901447-b91d-4b95-861d-8e6caf7aa121"). InnerVolumeSpecName "kube-api-access-gtblk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.237612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2" (OuterVolumeSpecName: "kube-api-access-4lmv2") pod "65475df3-6ce0-44d3-a5ed-861e34334058" (UID: "65475df3-6ce0-44d3-a5ed-861e34334058"). InnerVolumeSpecName "kube-api-access-4lmv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.249886 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65475df3-6ce0-44d3-a5ed-861e34334058" (UID: "65475df3-6ce0-44d3-a5ed-861e34334058"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.264189 4960 scope.go:117] "RemoveContainer" containerID="a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.282230 4960 scope.go:117] "RemoveContainer" containerID="f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.282621 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be\": container with ID starting with f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be not found: ID does not exist" containerID="f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.282670 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be"} err="failed to get container status \"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be\": rpc error: code = NotFound desc = could not find container \"f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be\": container with ID starting with f5eff7ad0a2fc9ab48463324eae0b58a168cdba746cd768cf33563f38077b8be not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.282701 4960 scope.go:117] "RemoveContainer" containerID="7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.283193 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1\": container with ID starting with 7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1 not found: ID does not exist" containerID="7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.283220 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1"} err="failed to get container status \"7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1\": rpc error: code = NotFound desc = could not find container \"7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1\": container with ID starting with 7af18ed24b9531ac2660a69d4c0a43a51eed9a1f32e8ae6a17c1d7da3d33d5b1 not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.283239 4960 scope.go:117] "RemoveContainer" containerID="a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05" Dec 01 19:45:05 crc kubenswrapper[4960]: E1201 19:45:05.283644 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05\": container with ID starting with a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05 not found: ID does not exist" containerID="a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.283668 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05"} err="failed to get container status \"a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05\": rpc error: code = NotFound desc = could not find container \"a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05\": container with ID starting with a8abab724c22cea1e1158e37dc2077a239d468ad32bda111189e61a124f10e05 not found: ID does not exist" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.283686 4960 scope.go:117] "RemoveContainer" containerID="90d2540886ef78985b81521d66781c786eb822df3aa3ec28b19ea93c432d7624" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.292065 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3901447-b91d-4b95-861d-8e6caf7aa121" (UID: "d3901447-b91d-4b95-861d-8e6caf7aa121"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331039 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331251 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65475df3-6ce0-44d3-a5ed-861e34334058-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331343 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lmv2\" (UniqueName: \"kubernetes.io/projected/65475df3-6ce0-44d3-a5ed-861e34334058-kube-api-access-4lmv2\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331424 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtblk\" (UniqueName: \"kubernetes.io/projected/d3901447-b91d-4b95-861d-8e6caf7aa121-kube-api-access-gtblk\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331500 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.331584 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3901447-b91d-4b95-861d-8e6caf7aa121-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.373881 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:45:05 crc kubenswrapper[4960]: W1201 19:45:05.378287 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod544fa4c7_f389_4765_8880_d6fa67065167.slice/crio-f56e66b2f130446730f77050f680bc879f0aca7419f7ebcf4140fc1b58e4b85f WatchSource:0}: Error finding container f56e66b2f130446730f77050f680bc879f0aca7419f7ebcf4140fc1b58e4b85f: Status 404 returned error can't find the container with id f56e66b2f130446730f77050f680bc879f0aca7419f7ebcf4140fc1b58e4b85f Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.486372 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.490952 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7ldv"] Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.502328 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.503041 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jjmnx"] Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.589284 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.635464 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mndrf\" (UniqueName: \"kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf\") pod \"d723bc3b-510d-46fb-999b-6d0b03ddb349\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.635545 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics\") pod \"d723bc3b-510d-46fb-999b-6d0b03ddb349\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.643442 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf" (OuterVolumeSpecName: "kube-api-access-mndrf") pod "d723bc3b-510d-46fb-999b-6d0b03ddb349" (UID: "d723bc3b-510d-46fb-999b-6d0b03ddb349"). InnerVolumeSpecName "kube-api-access-mndrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.643516 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "d723bc3b-510d-46fb-999b-6d0b03ddb349" (UID: "d723bc3b-510d-46fb-999b-6d0b03ddb349"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.688346 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.692005 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736547 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2lqt\" (UniqueName: \"kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt\") pod \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736611 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities\") pod \"3c55929c-b18d-4719-944e-54176bdb96bb\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736671 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4wvs\" (UniqueName: \"kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs\") pod \"3c55929c-b18d-4719-944e-54176bdb96bb\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736730 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca\") pod \"d723bc3b-510d-46fb-999b-6d0b03ddb349\" (UID: \"d723bc3b-510d-46fb-999b-6d0b03ddb349\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content\") pod \"3c55929c-b18d-4719-944e-54176bdb96bb\" (UID: \"3c55929c-b18d-4719-944e-54176bdb96bb\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content\") pod \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.736935 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities\") pod \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\" (UID: \"9f2e47de-2360-4506-9e43-ee29da7ca9e1\") " Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.737234 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mndrf\" (UniqueName: \"kubernetes.io/projected/d723bc3b-510d-46fb-999b-6d0b03ddb349-kube-api-access-mndrf\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.737248 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.737502 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities" (OuterVolumeSpecName: "utilities") pod "3c55929c-b18d-4719-944e-54176bdb96bb" (UID: "3c55929c-b18d-4719-944e-54176bdb96bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.738376 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "d723bc3b-510d-46fb-999b-6d0b03ddb349" (UID: "d723bc3b-510d-46fb-999b-6d0b03ddb349"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.740556 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs" (OuterVolumeSpecName: "kube-api-access-h4wvs") pod "3c55929c-b18d-4719-944e-54176bdb96bb" (UID: "3c55929c-b18d-4719-944e-54176bdb96bb"). InnerVolumeSpecName "kube-api-access-h4wvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.741700 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities" (OuterVolumeSpecName: "utilities") pod "9f2e47de-2360-4506-9e43-ee29da7ca9e1" (UID: "9f2e47de-2360-4506-9e43-ee29da7ca9e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.746770 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt" (OuterVolumeSpecName: "kube-api-access-j2lqt") pod "9f2e47de-2360-4506-9e43-ee29da7ca9e1" (UID: "9f2e47de-2360-4506-9e43-ee29da7ca9e1"). InnerVolumeSpecName "kube-api-access-j2lqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.818296 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f2e47de-2360-4506-9e43-ee29da7ca9e1" (UID: "9f2e47de-2360-4506-9e43-ee29da7ca9e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839089 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839588 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2e47de-2360-4506-9e43-ee29da7ca9e1-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839612 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2lqt\" (UniqueName: \"kubernetes.io/projected/9f2e47de-2360-4506-9e43-ee29da7ca9e1-kube-api-access-j2lqt\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839627 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839640 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4wvs\" (UniqueName: \"kubernetes.io/projected/3c55929c-b18d-4719-944e-54176bdb96bb-kube-api-access-h4wvs\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.839652 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d723bc3b-510d-46fb-999b-6d0b03ddb349-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.880070 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c55929c-b18d-4719-944e-54176bdb96bb" (UID: "3c55929c-b18d-4719-944e-54176bdb96bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:05 crc kubenswrapper[4960]: I1201 19:45:05.941135 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c55929c-b18d-4719-944e-54176bdb96bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.205268 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" event={"ID":"d723bc3b-510d-46fb-999b-6d0b03ddb349","Type":"ContainerDied","Data":"c802b1029eecf5869c1561db4a2a934a9dc274e5e20550682333f1a2947b1106"} Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.205333 4960 scope.go:117] "RemoveContainer" containerID="0201347d563154d6d9b37e714f6edbf824617b5930fd30d27ebe6a61be27addb" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.205363 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w6wcz" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.213451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlm2j" event={"ID":"9f2e47de-2360-4506-9e43-ee29da7ca9e1","Type":"ContainerDied","Data":"b0d3eef713b4fe4ae19b780413a34a14e98e30cef7a4d456c0c5538579f3822f"} Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.213590 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlm2j" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.219864 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwhqw" event={"ID":"3c55929c-b18d-4719-944e-54176bdb96bb","Type":"ContainerDied","Data":"244d51c5f070ed3060757729c980afa0eb2e8bb66f8c1e3201daad725cd3ca0f"} Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.220019 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwhqw" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.225800 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" event={"ID":"544fa4c7-f389-4765-8880-d6fa67065167","Type":"ContainerStarted","Data":"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000"} Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.226882 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" event={"ID":"544fa4c7-f389-4765-8880-d6fa67065167","Type":"ContainerStarted","Data":"f56e66b2f130446730f77050f680bc879f0aca7419f7ebcf4140fc1b58e4b85f"} Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.227037 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.235481 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.243237 4960 scope.go:117] "RemoveContainer" containerID="d775fcaaba693f9fcaa8fe641447a1c290eac7df75a46727b6868911028329c7" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.256221 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" podStartSLOduration=2.256192916 podStartE2EDuration="2.256192916s" podCreationTimestamp="2025-12-01 19:45:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:45:06.247938843 +0000 UTC m=+341.535430572" watchObservedRunningTime="2025-12-01 19:45:06.256192916 +0000 UTC m=+341.543684615" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.271755 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.274431 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zlm2j"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.284572 4960 scope.go:117] "RemoveContainer" containerID="999259a8921729ec0761ae460f414646bd0956547f14e52356c4494dc7803c5c" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.287234 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.294861 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fwhqw"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.306374 4960 scope.go:117] "RemoveContainer" containerID="041dafbd13346d3fde0f7142edebb895870e1ff9d757798c345cd6875c8415b3" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.308912 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.321168 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w6wcz"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.347137 4960 scope.go:117] "RemoveContainer" containerID="6046f1ee63a0db3d6160afbed536e144f905bf9a2deb52fcd2fcffda1837a51a" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.361964 4960 scope.go:117] "RemoveContainer" containerID="bc37ae77178bbd76e5cf1cdfacf6c1d8559607cec588b291cabc38725cce6968" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.375885 4960 scope.go:117] "RemoveContainer" containerID="146dec8baf84c5bf60700f546db42ca1c022600e6abb0af144e426bf2d7affc3" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879275 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879721 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879751 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879772 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879790 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879817 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879833 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879857 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879873 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879889 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879904 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879925 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879940 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.879967 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.879982 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880006 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880022 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880087 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880106 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="extract-content" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880165 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880182 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880210 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880225 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880241 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880255 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880281 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880297 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="extract-utilities" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880501 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880526 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880550 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880573 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880597 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880616 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" containerName="registry-server" Dec 01 19:45:06 crc kubenswrapper[4960]: E1201 19:45:06.880853 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.880874 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" containerName="marketplace-operator" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.882376 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.886338 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.900328 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.956074 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzdb2\" (UniqueName: \"kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.956403 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:06 crc kubenswrapper[4960]: I1201 19:45:06.956472 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.057835 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzdb2\" (UniqueName: \"kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.057905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.057947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.059218 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.060048 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.083792 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzdb2\" (UniqueName: \"kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2\") pod \"community-operators-xct8q\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.205201 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.335477 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c55929c-b18d-4719-944e-54176bdb96bb" path="/var/lib/kubelet/pods/3c55929c-b18d-4719-944e-54176bdb96bb/volumes" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.336506 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65475df3-6ce0-44d3-a5ed-861e34334058" path="/var/lib/kubelet/pods/65475df3-6ce0-44d3-a5ed-861e34334058/volumes" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.337238 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2e47de-2360-4506-9e43-ee29da7ca9e1" path="/var/lib/kubelet/pods/9f2e47de-2360-4506-9e43-ee29da7ca9e1/volumes" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.338535 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3901447-b91d-4b95-861d-8e6caf7aa121" path="/var/lib/kubelet/pods/d3901447-b91d-4b95-861d-8e6caf7aa121/volumes" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.339381 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d723bc3b-510d-46fb-999b-6d0b03ddb349" path="/var/lib/kubelet/pods/d723bc3b-510d-46fb-999b-6d0b03ddb349/volumes" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.652000 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:45:07 crc kubenswrapper[4960]: W1201 19:45:07.661265 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b458ace_89d3_4a6f_950e_a7f54f90f340.slice/crio-cca32d5375867dbdf3fd0ec2862c1658ce3f23783e1c1bcde9e863b1587cec8d WatchSource:0}: Error finding container cca32d5375867dbdf3fd0ec2862c1658ce3f23783e1c1bcde9e863b1587cec8d: Status 404 returned error can't find the container with id cca32d5375867dbdf3fd0ec2862c1658ce3f23783e1c1bcde9e863b1587cec8d Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.786485 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rztn8"] Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.787705 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.803162 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rztn8"] Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.969837 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-bound-sa-token\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.969920 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccx2d\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-kube-api-access-ccx2d\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.969957 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.970142 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-tls\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.970225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-trusted-ca\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.970781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-certificates\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.970919 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e436c6a-d616-48ae-a74b-9400772d4d0c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.971106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e436c6a-d616-48ae-a74b-9400772d4d0c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:07 crc kubenswrapper[4960]: I1201 19:45:07.992033 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072323 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccx2d\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-kube-api-access-ccx2d\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072493 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-tls\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072538 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-trusted-ca\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072591 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-certificates\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072646 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e436c6a-d616-48ae-a74b-9400772d4d0c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072749 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e436c6a-d616-48ae-a74b-9400772d4d0c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.072799 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-bound-sa-token\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.073740 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2e436c6a-d616-48ae-a74b-9400772d4d0c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.075093 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-certificates\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.075865 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e436c6a-d616-48ae-a74b-9400772d4d0c-trusted-ca\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.083243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-registry-tls\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.084669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2e436c6a-d616-48ae-a74b-9400772d4d0c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.089949 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccx2d\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-kube-api-access-ccx2d\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.099014 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2e436c6a-d616-48ae-a74b-9400772d4d0c-bound-sa-token\") pod \"image-registry-66df7c8f76-rztn8\" (UID: \"2e436c6a-d616-48ae-a74b-9400772d4d0c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.102961 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.250476 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerID="acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54" exitCode=0 Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.250608 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerDied","Data":"acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54"} Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.250977 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerStarted","Data":"cca32d5375867dbdf3fd0ec2862c1658ce3f23783e1c1bcde9e863b1587cec8d"} Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.272584 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.273924 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.284166 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.300972 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.376745 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.376797 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.377147 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp8zf\" (UniqueName: \"kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.479163 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp8zf\" (UniqueName: \"kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.479242 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.479265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.479854 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.479961 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.499050 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp8zf\" (UniqueName: \"kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf\") pod \"redhat-operators-qwd6x\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.574319 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rztn8"] Dec 01 19:45:08 crc kubenswrapper[4960]: I1201 19:45:08.607540 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.043377 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:45:09 crc kubenswrapper[4960]: W1201 19:45:09.052085 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15c5f8a1_4662_457f_962a_c7ba801056e8.slice/crio-45a1f2fab9838c94bf25104e0d2032ab99c941fcae0cb6e3758e7330784640d4 WatchSource:0}: Error finding container 45a1f2fab9838c94bf25104e0d2032ab99c941fcae0cb6e3758e7330784640d4: Status 404 returned error can't find the container with id 45a1f2fab9838c94bf25104e0d2032ab99c941fcae0cb6e3758e7330784640d4 Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.263996 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.266225 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.268904 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.271005 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" event={"ID":"2e436c6a-d616-48ae-a74b-9400772d4d0c","Type":"ContainerStarted","Data":"cc6f81fbc4c70031216529b307d33dfd6efe6f5a5216bd3575e91b65effa76db"} Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.271094 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.271124 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" event={"ID":"2e436c6a-d616-48ae-a74b-9400772d4d0c","Type":"ContainerStarted","Data":"0d5d724c8e739caf38c99d2e0575ae3adab7a17786289a6746ca2032706059f9"} Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.273530 4960 generic.go:334] "Generic (PLEG): container finished" podID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerID="eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df" exitCode=0 Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.273610 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerDied","Data":"eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df"} Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.273658 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerStarted","Data":"45a1f2fab9838c94bf25104e0d2032ab99c941fcae0cb6e3758e7330784640d4"} Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.286236 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.294098 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.294467 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.294545 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8p55\" (UniqueName: \"kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.335141 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" podStartSLOduration=2.335094737 podStartE2EDuration="2.335094737s" podCreationTimestamp="2025-12-01 19:45:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:45:09.330683886 +0000 UTC m=+344.618175575" watchObservedRunningTime="2025-12-01 19:45:09.335094737 +0000 UTC m=+344.622586406" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.395731 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.396274 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.396296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8p55\" (UniqueName: \"kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.397251 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.397572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.428980 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8p55\" (UniqueName: \"kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55\") pod \"redhat-marketplace-69sk4\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:09 crc kubenswrapper[4960]: I1201 19:45:09.641217 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.072919 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:45:10 crc kubenswrapper[4960]: W1201 19:45:10.085390 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05309d39_9cfb_430f_b69f_9ace1978796f.slice/crio-64af68bbb5cfd59872cd424d3da698e86993d6916ea97b2cde83c6c1c043193f WatchSource:0}: Error finding container 64af68bbb5cfd59872cd424d3da698e86993d6916ea97b2cde83c6c1c043193f: Status 404 returned error can't find the container with id 64af68bbb5cfd59872cd424d3da698e86993d6916ea97b2cde83c6c1c043193f Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.116498 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.116719 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" podUID="24d71258-296b-4a3c-b54c-ada2e334f9ca" containerName="route-controller-manager" containerID="cri-o://8474b4a7f671fe29631cb74fea4caa7a1e44f2d956fdafdbe5129c0b8ca10bb1" gracePeriod=30 Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.280865 4960 generic.go:334] "Generic (PLEG): container finished" podID="24d71258-296b-4a3c-b54c-ada2e334f9ca" containerID="8474b4a7f671fe29631cb74fea4caa7a1e44f2d956fdafdbe5129c0b8ca10bb1" exitCode=0 Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.281050 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" event={"ID":"24d71258-296b-4a3c-b54c-ada2e334f9ca","Type":"ContainerDied","Data":"8474b4a7f671fe29631cb74fea4caa7a1e44f2d956fdafdbe5129c0b8ca10bb1"} Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.284257 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerID="0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd" exitCode=0 Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.284330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerDied","Data":"0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd"} Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.285331 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerStarted","Data":"64af68bbb5cfd59872cd424d3da698e86993d6916ea97b2cde83c6c1c043193f"} Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.484279 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.613694 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert\") pod \"24d71258-296b-4a3c-b54c-ada2e334f9ca\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.613790 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config\") pod \"24d71258-296b-4a3c-b54c-ada2e334f9ca\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.613815 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca\") pod \"24d71258-296b-4a3c-b54c-ada2e334f9ca\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.613922 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j75kz\" (UniqueName: \"kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz\") pod \"24d71258-296b-4a3c-b54c-ada2e334f9ca\" (UID: \"24d71258-296b-4a3c-b54c-ada2e334f9ca\") " Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.615177 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca" (OuterVolumeSpecName: "client-ca") pod "24d71258-296b-4a3c-b54c-ada2e334f9ca" (UID: "24d71258-296b-4a3c-b54c-ada2e334f9ca"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.615378 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config" (OuterVolumeSpecName: "config") pod "24d71258-296b-4a3c-b54c-ada2e334f9ca" (UID: "24d71258-296b-4a3c-b54c-ada2e334f9ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.620017 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz" (OuterVolumeSpecName: "kube-api-access-j75kz") pod "24d71258-296b-4a3c-b54c-ada2e334f9ca" (UID: "24d71258-296b-4a3c-b54c-ada2e334f9ca"). InnerVolumeSpecName "kube-api-access-j75kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.620018 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "24d71258-296b-4a3c-b54c-ada2e334f9ca" (UID: "24d71258-296b-4a3c-b54c-ada2e334f9ca"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.682100 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:45:10 crc kubenswrapper[4960]: E1201 19:45:10.682418 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d71258-296b-4a3c-b54c-ada2e334f9ca" containerName="route-controller-manager" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.682436 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d71258-296b-4a3c-b54c-ada2e334f9ca" containerName="route-controller-manager" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.682545 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d71258-296b-4a3c-b54c-ada2e334f9ca" containerName="route-controller-manager" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.683554 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.686862 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.687189 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.715367 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.715436 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/24d71258-296b-4a3c-b54c-ada2e334f9ca-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.715454 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j75kz\" (UniqueName: \"kubernetes.io/projected/24d71258-296b-4a3c-b54c-ada2e334f9ca-kube-api-access-j75kz\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.715469 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/24d71258-296b-4a3c-b54c-ada2e334f9ca-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.816844 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjbnf\" (UniqueName: \"kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.816900 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.816923 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.917969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjbnf\" (UniqueName: \"kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.918437 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.918460 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.918985 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.919234 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:10 crc kubenswrapper[4960]: I1201 19:45:10.943783 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjbnf\" (UniqueName: \"kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf\") pod \"certified-operators-vhhcs\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.003683 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.293484 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" event={"ID":"24d71258-296b-4a3c-b54c-ada2e334f9ca","Type":"ContainerDied","Data":"84df0853123fe2373731bb067a8b4eee1997c76a8bc2100ddf1052261862a6ac"} Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.293533 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.293570 4960 scope.go:117] "RemoveContainer" containerID="8474b4a7f671fe29631cb74fea4caa7a1e44f2d956fdafdbe5129c0b8ca10bb1" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.297082 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerStarted","Data":"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7"} Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.301688 4960 generic.go:334] "Generic (PLEG): container finished" podID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerID="683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60" exitCode=0 Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.301756 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerDied","Data":"683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60"} Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.305714 4960 generic.go:334] "Generic (PLEG): container finished" podID="05309d39-9cfb-430f-b69f-9ace1978796f" containerID="9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172" exitCode=0 Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.305766 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerDied","Data":"9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172"} Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.316885 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xct8q" podStartSLOduration=2.828255904 podStartE2EDuration="5.316867313s" podCreationTimestamp="2025-12-01 19:45:06 +0000 UTC" firstStartedPulling="2025-12-01 19:45:08.252512568 +0000 UTC m=+343.540004237" lastFinishedPulling="2025-12-01 19:45:10.741123977 +0000 UTC m=+346.028615646" observedRunningTime="2025-12-01 19:45:11.314960022 +0000 UTC m=+346.602451701" watchObservedRunningTime="2025-12-01 19:45:11.316867313 +0000 UTC m=+346.604358982" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.374479 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.377659 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6ccb45c985-lwqlh"] Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.423543 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:45:11 crc kubenswrapper[4960]: W1201 19:45:11.434932 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0237c051_807d_4c6c_8f72_71cd35f2a8df.slice/crio-7fbd601fcba81ec6483beb340f211e946e742bd3ca05ac43406a8329b52fba99 WatchSource:0}: Error finding container 7fbd601fcba81ec6483beb340f211e946e742bd3ca05ac43406a8329b52fba99: Status 404 returned error can't find the container with id 7fbd601fcba81ec6483beb340f211e946e742bd3ca05ac43406a8329b52fba99 Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.830265 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh"] Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.831498 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.834463 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.834492 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.834712 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.838858 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.839125 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.840389 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.866398 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh"] Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.932827 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-config\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.933181 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-client-ca\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.933316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ddb9ed-6def-4022-b400-e073c6d7db30-serving-cert\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:11 crc kubenswrapper[4960]: I1201 19:45:11.933493 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wv9s\" (UniqueName: \"kubernetes.io/projected/a7ddb9ed-6def-4022-b400-e073c6d7db30-kube-api-access-4wv9s\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.035306 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wv9s\" (UniqueName: \"kubernetes.io/projected/a7ddb9ed-6def-4022-b400-e073c6d7db30-kube-api-access-4wv9s\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.035371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-config\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.035401 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-client-ca\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.035425 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ddb9ed-6def-4022-b400-e073c6d7db30-serving-cert\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.036727 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-client-ca\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.037282 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7ddb9ed-6def-4022-b400-e073c6d7db30-config\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.043033 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ddb9ed-6def-4022-b400-e073c6d7db30-serving-cert\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.055819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wv9s\" (UniqueName: \"kubernetes.io/projected/a7ddb9ed-6def-4022-b400-e073c6d7db30-kube-api-access-4wv9s\") pod \"route-controller-manager-55486b59b5-j2bqh\" (UID: \"a7ddb9ed-6def-4022-b400-e073c6d7db30\") " pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.149832 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.315659 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerStarted","Data":"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e"} Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.321094 4960 generic.go:334] "Generic (PLEG): container finished" podID="05309d39-9cfb-430f-b69f-9ace1978796f" containerID="a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549" exitCode=0 Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.321155 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerDied","Data":"a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549"} Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.354379 4960 generic.go:334] "Generic (PLEG): container finished" podID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerID="6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1" exitCode=0 Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.354735 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerDied","Data":"6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1"} Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.354792 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerStarted","Data":"7fbd601fcba81ec6483beb340f211e946e742bd3ca05ac43406a8329b52fba99"} Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.364146 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qwd6x" podStartSLOduration=1.531636179 podStartE2EDuration="4.364106373s" podCreationTimestamp="2025-12-01 19:45:08 +0000 UTC" firstStartedPulling="2025-12-01 19:45:09.276021678 +0000 UTC m=+344.563513377" lastFinishedPulling="2025-12-01 19:45:12.108491902 +0000 UTC m=+347.395983571" observedRunningTime="2025-12-01 19:45:12.339758465 +0000 UTC m=+347.627250134" watchObservedRunningTime="2025-12-01 19:45:12.364106373 +0000 UTC m=+347.651598032" Dec 01 19:45:12 crc kubenswrapper[4960]: W1201 19:45:12.562141 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7ddb9ed_6def_4022_b400_e073c6d7db30.slice/crio-75b13c87ef59e3e29717c18b5387a215340d0d8aa4dc5aa7248d7d0e72c58e43 WatchSource:0}: Error finding container 75b13c87ef59e3e29717c18b5387a215340d0d8aa4dc5aa7248d7d0e72c58e43: Status 404 returned error can't find the container with id 75b13c87ef59e3e29717c18b5387a215340d0d8aa4dc5aa7248d7d0e72c58e43 Dec 01 19:45:12 crc kubenswrapper[4960]: I1201 19:45:12.562725 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh"] Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.334572 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d71258-296b-4a3c-b54c-ada2e334f9ca" path="/var/lib/kubelet/pods/24d71258-296b-4a3c-b54c-ada2e334f9ca/volumes" Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.362458 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerStarted","Data":"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778"} Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.364084 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" event={"ID":"a7ddb9ed-6def-4022-b400-e073c6d7db30","Type":"ContainerStarted","Data":"34e09c3e3659d8167c39ab75e97740a76814f71a53ff2e149bd5dc85293a10eb"} Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.364161 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" event={"ID":"a7ddb9ed-6def-4022-b400-e073c6d7db30","Type":"ContainerStarted","Data":"75b13c87ef59e3e29717c18b5387a215340d0d8aa4dc5aa7248d7d0e72c58e43"} Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.364298 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.367067 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerStarted","Data":"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324"} Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.372629 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.400905 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55486b59b5-j2bqh" podStartSLOduration=3.400882898 podStartE2EDuration="3.400882898s" podCreationTimestamp="2025-12-01 19:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:45:13.398228503 +0000 UTC m=+348.685720172" watchObservedRunningTime="2025-12-01 19:45:13.400882898 +0000 UTC m=+348.688374567" Dec 01 19:45:13 crc kubenswrapper[4960]: I1201 19:45:13.416438 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-69sk4" podStartSLOduration=2.919310894 podStartE2EDuration="4.416411495s" podCreationTimestamp="2025-12-01 19:45:09 +0000 UTC" firstStartedPulling="2025-12-01 19:45:11.306935956 +0000 UTC m=+346.594427625" lastFinishedPulling="2025-12-01 19:45:12.804036547 +0000 UTC m=+348.091528226" observedRunningTime="2025-12-01 19:45:13.415429433 +0000 UTC m=+348.702921102" watchObservedRunningTime="2025-12-01 19:45:13.416411495 +0000 UTC m=+348.703903174" Dec 01 19:45:14 crc kubenswrapper[4960]: I1201 19:45:14.378851 4960 generic.go:334] "Generic (PLEG): container finished" podID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerID="75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778" exitCode=0 Dec 01 19:45:14 crc kubenswrapper[4960]: I1201 19:45:14.378956 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerDied","Data":"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778"} Dec 01 19:45:16 crc kubenswrapper[4960]: I1201 19:45:16.396355 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerStarted","Data":"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80"} Dec 01 19:45:16 crc kubenswrapper[4960]: I1201 19:45:16.416722 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vhhcs" podStartSLOduration=3.084986078 podStartE2EDuration="6.416699732s" podCreationTimestamp="2025-12-01 19:45:10 +0000 UTC" firstStartedPulling="2025-12-01 19:45:12.357169991 +0000 UTC m=+347.644661650" lastFinishedPulling="2025-12-01 19:45:15.688883635 +0000 UTC m=+350.976375304" observedRunningTime="2025-12-01 19:45:16.413376016 +0000 UTC m=+351.700867685" watchObservedRunningTime="2025-12-01 19:45:16.416699732 +0000 UTC m=+351.704191401" Dec 01 19:45:17 crc kubenswrapper[4960]: I1201 19:45:17.206136 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:17 crc kubenswrapper[4960]: I1201 19:45:17.206250 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:17 crc kubenswrapper[4960]: I1201 19:45:17.253798 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:17 crc kubenswrapper[4960]: I1201 19:45:17.455888 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:45:18 crc kubenswrapper[4960]: I1201 19:45:18.608647 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:18 crc kubenswrapper[4960]: I1201 19:45:18.609161 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:18 crc kubenswrapper[4960]: I1201 19:45:18.657775 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:19 crc kubenswrapper[4960]: I1201 19:45:19.456981 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:45:19 crc kubenswrapper[4960]: I1201 19:45:19.642379 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:19 crc kubenswrapper[4960]: I1201 19:45:19.642472 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:19 crc kubenswrapper[4960]: I1201 19:45:19.701147 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:20 crc kubenswrapper[4960]: I1201 19:45:20.474492 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:45:21 crc kubenswrapper[4960]: I1201 19:45:21.006100 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:21 crc kubenswrapper[4960]: I1201 19:45:21.006489 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:21 crc kubenswrapper[4960]: I1201 19:45:21.047521 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:21 crc kubenswrapper[4960]: I1201 19:45:21.465973 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:45:28 crc kubenswrapper[4960]: I1201 19:45:28.110240 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rztn8" Dec 01 19:45:28 crc kubenswrapper[4960]: I1201 19:45:28.175715 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.109027 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.110629 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" podUID="d15ad974-a4c4-4a77-9a4a-470f34e53431" containerName="controller-manager" containerID="cri-o://20b095bf7fda0ca7346692005f8329d69d71bf9677e4825beb65e368e1bf802f" gracePeriod=30 Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.482708 4960 generic.go:334] "Generic (PLEG): container finished" podID="d15ad974-a4c4-4a77-9a4a-470f34e53431" containerID="20b095bf7fda0ca7346692005f8329d69d71bf9677e4825beb65e368e1bf802f" exitCode=0 Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.482777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" event={"ID":"d15ad974-a4c4-4a77-9a4a-470f34e53431","Type":"ContainerDied","Data":"20b095bf7fda0ca7346692005f8329d69d71bf9677e4825beb65e368e1bf802f"} Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.482810 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" event={"ID":"d15ad974-a4c4-4a77-9a4a-470f34e53431","Type":"ContainerDied","Data":"8ad7458c32e9fda7041b57cf1262184e21e458195cf9634f03d8ef739915e0f8"} Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.482822 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ad7458c32e9fda7041b57cf1262184e21e458195cf9634f03d8ef739915e0f8" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.525153 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.606269 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles\") pod \"d15ad974-a4c4-4a77-9a4a-470f34e53431\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.606330 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config\") pod \"d15ad974-a4c4-4a77-9a4a-470f34e53431\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.606417 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert\") pod \"d15ad974-a4c4-4a77-9a4a-470f34e53431\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.606513 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca\") pod \"d15ad974-a4c4-4a77-9a4a-470f34e53431\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.606570 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd2ps\" (UniqueName: \"kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps\") pod \"d15ad974-a4c4-4a77-9a4a-470f34e53431\" (UID: \"d15ad974-a4c4-4a77-9a4a-470f34e53431\") " Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.607623 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d15ad974-a4c4-4a77-9a4a-470f34e53431" (UID: "d15ad974-a4c4-4a77-9a4a-470f34e53431"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.608590 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca" (OuterVolumeSpecName: "client-ca") pod "d15ad974-a4c4-4a77-9a4a-470f34e53431" (UID: "d15ad974-a4c4-4a77-9a4a-470f34e53431"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.609034 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config" (OuterVolumeSpecName: "config") pod "d15ad974-a4c4-4a77-9a4a-470f34e53431" (UID: "d15ad974-a4c4-4a77-9a4a-470f34e53431"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.614689 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d15ad974-a4c4-4a77-9a4a-470f34e53431" (UID: "d15ad974-a4c4-4a77-9a4a-470f34e53431"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.618355 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps" (OuterVolumeSpecName: "kube-api-access-nd2ps") pod "d15ad974-a4c4-4a77-9a4a-470f34e53431" (UID: "d15ad974-a4c4-4a77-9a4a-470f34e53431"). InnerVolumeSpecName "kube-api-access-nd2ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.708001 4960 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.708049 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.708061 4960 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d15ad974-a4c4-4a77-9a4a-470f34e53431-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.708070 4960 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d15ad974-a4c4-4a77-9a4a-470f34e53431-client-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:30 crc kubenswrapper[4960]: I1201 19:45:30.708082 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd2ps\" (UniqueName: \"kubernetes.io/projected/d15ad974-a4c4-4a77-9a4a-470f34e53431-kube-api-access-nd2ps\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.488591 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.512026 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.518273 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7766fdf5d9-7nrd8"] Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.852708 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5cf5874cf4-mm25m"] Dec 01 19:45:31 crc kubenswrapper[4960]: E1201 19:45:31.853034 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15ad974-a4c4-4a77-9a4a-470f34e53431" containerName="controller-manager" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.853055 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15ad974-a4c4-4a77-9a4a-470f34e53431" containerName="controller-manager" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.853261 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15ad974-a4c4-4a77-9a4a-470f34e53431" containerName="controller-manager" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.853877 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.856639 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.856960 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.857172 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.858957 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.858969 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.863699 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cf5874cf4-mm25m"] Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.865744 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.868164 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.923464 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfm9r\" (UniqueName: \"kubernetes.io/projected/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-kube-api-access-pfm9r\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.923632 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-config\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.923802 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-serving-cert\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.923847 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-client-ca\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:31 crc kubenswrapper[4960]: I1201 19:45:31.923933 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-proxy-ca-bundles\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.025652 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-proxy-ca-bundles\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.025738 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfm9r\" (UniqueName: \"kubernetes.io/projected/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-kube-api-access-pfm9r\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.025775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-config\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.025804 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-serving-cert\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.025819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-client-ca\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.027274 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-client-ca\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.028012 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-proxy-ca-bundles\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.028810 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-config\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.031788 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-serving-cert\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.043185 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfm9r\" (UniqueName: \"kubernetes.io/projected/a292a80d-ce9b-40f1-87ed-e67b249ee6ac-kube-api-access-pfm9r\") pod \"controller-manager-5cf5874cf4-mm25m\" (UID: \"a292a80d-ce9b-40f1-87ed-e67b249ee6ac\") " pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.174773 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:32 crc kubenswrapper[4960]: I1201 19:45:32.611382 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cf5874cf4-mm25m"] Dec 01 19:45:32 crc kubenswrapper[4960]: W1201 19:45:32.623881 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda292a80d_ce9b_40f1_87ed_e67b249ee6ac.slice/crio-59f52ca84c04ba955e073000ecc463103b638126bc3829e960bade79c156301e WatchSource:0}: Error finding container 59f52ca84c04ba955e073000ecc463103b638126bc3829e960bade79c156301e: Status 404 returned error can't find the container with id 59f52ca84c04ba955e073000ecc463103b638126bc3829e960bade79c156301e Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.331818 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15ad974-a4c4-4a77-9a4a-470f34e53431" path="/var/lib/kubelet/pods/d15ad974-a4c4-4a77-9a4a-470f34e53431/volumes" Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.501166 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" event={"ID":"a292a80d-ce9b-40f1-87ed-e67b249ee6ac","Type":"ContainerStarted","Data":"250a532f56649bbde3a7bcb0abe26bcc520169a8f38593c080c9275b2ee89bfb"} Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.501219 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" event={"ID":"a292a80d-ce9b-40f1-87ed-e67b249ee6ac","Type":"ContainerStarted","Data":"59f52ca84c04ba955e073000ecc463103b638126bc3829e960bade79c156301e"} Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.501819 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.507334 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" Dec 01 19:45:33 crc kubenswrapper[4960]: I1201 19:45:33.544435 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5cf5874cf4-mm25m" podStartSLOduration=3.544407096 podStartE2EDuration="3.544407096s" podCreationTimestamp="2025-12-01 19:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:45:33.522870848 +0000 UTC m=+368.810362527" watchObservedRunningTime="2025-12-01 19:45:33.544407096 +0000 UTC m=+368.831898775" Dec 01 19:45:43 crc kubenswrapper[4960]: I1201 19:45:43.870797 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:45:43 crc kubenswrapper[4960]: I1201 19:45:43.871832 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.216818 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" podUID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" containerName="registry" containerID="cri-o://faedc782190686d1686a51a5032b561d0ac9c718712d35b4a33c5115b051f82a" gracePeriod=30 Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.670379 4960 generic.go:334] "Generic (PLEG): container finished" podID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" containerID="faedc782190686d1686a51a5032b561d0ac9c718712d35b4a33c5115b051f82a" exitCode=0 Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.670624 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" event={"ID":"41dd0037-d130-4fcf-8507-4dfa3059a9d9","Type":"ContainerDied","Data":"faedc782190686d1686a51a5032b561d0ac9c718712d35b4a33c5115b051f82a"} Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.760802 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.889871 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.889949 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890002 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890243 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890295 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890343 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.890422 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh2kr\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr\") pod \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\" (UID: \"41dd0037-d130-4fcf-8507-4dfa3059a9d9\") " Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.891395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.891931 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.898787 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.899195 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.899557 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.900284 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr" (OuterVolumeSpecName: "kube-api-access-wh2kr") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "kube-api-access-wh2kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.909941 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.921071 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "41dd0037-d130-4fcf-8507-4dfa3059a9d9" (UID: "41dd0037-d130-4fcf-8507-4dfa3059a9d9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991792 4960 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991854 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh2kr\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-kube-api-access-wh2kr\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991873 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/41dd0037-d130-4fcf-8507-4dfa3059a9d9-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991893 4960 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991910 4960 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/41dd0037-d130-4fcf-8507-4dfa3059a9d9-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991927 4960 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/41dd0037-d130-4fcf-8507-4dfa3059a9d9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:53 crc kubenswrapper[4960]: I1201 19:45:53.991944 4960 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/41dd0037-d130-4fcf-8507-4dfa3059a9d9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 01 19:45:54 crc kubenswrapper[4960]: I1201 19:45:54.678620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" event={"ID":"41dd0037-d130-4fcf-8507-4dfa3059a9d9","Type":"ContainerDied","Data":"42097c77da0911afbf7d83a8a1db301627769d32b84eac5b547f87ea687d4e33"} Dec 01 19:45:54 crc kubenswrapper[4960]: I1201 19:45:54.678688 4960 scope.go:117] "RemoveContainer" containerID="faedc782190686d1686a51a5032b561d0ac9c718712d35b4a33c5115b051f82a" Dec 01 19:45:54 crc kubenswrapper[4960]: I1201 19:45:54.678751 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ccbrv" Dec 01 19:45:54 crc kubenswrapper[4960]: I1201 19:45:54.729366 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:45:54 crc kubenswrapper[4960]: I1201 19:45:54.735757 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ccbrv"] Dec 01 19:45:55 crc kubenswrapper[4960]: I1201 19:45:55.338850 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" path="/var/lib/kubelet/pods/41dd0037-d130-4fcf-8507-4dfa3059a9d9/volumes" Dec 01 19:46:13 crc kubenswrapper[4960]: I1201 19:46:13.871330 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:46:13 crc kubenswrapper[4960]: I1201 19:46:13.872422 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:46:43 crc kubenswrapper[4960]: I1201 19:46:43.871325 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:46:43 crc kubenswrapper[4960]: I1201 19:46:43.872387 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:46:43 crc kubenswrapper[4960]: I1201 19:46:43.872441 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:46:43 crc kubenswrapper[4960]: I1201 19:46:43.873241 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 19:46:43 crc kubenswrapper[4960]: I1201 19:46:43.873524 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967" gracePeriod=600 Dec 01 19:46:44 crc kubenswrapper[4960]: I1201 19:46:44.047936 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967" exitCode=0 Dec 01 19:46:44 crc kubenswrapper[4960]: I1201 19:46:44.048014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967"} Dec 01 19:46:44 crc kubenswrapper[4960]: I1201 19:46:44.048076 4960 scope.go:117] "RemoveContainer" containerID="fbe6be553d5eece76e5c84619601893c33cb7995eb4ac0bdace8579bcb088545" Dec 01 19:46:45 crc kubenswrapper[4960]: I1201 19:46:45.061330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b"} Dec 01 19:49:13 crc kubenswrapper[4960]: I1201 19:49:13.871450 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:49:13 crc kubenswrapper[4960]: I1201 19:49:13.872682 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:49:43 crc kubenswrapper[4960]: I1201 19:49:43.870472 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:49:43 crc kubenswrapper[4960]: I1201 19:49:43.871514 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:50:13 crc kubenswrapper[4960]: I1201 19:50:13.870550 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:50:13 crc kubenswrapper[4960]: I1201 19:50:13.871802 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:50:13 crc kubenswrapper[4960]: I1201 19:50:13.871867 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:50:13 crc kubenswrapper[4960]: I1201 19:50:13.872801 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 19:50:13 crc kubenswrapper[4960]: I1201 19:50:13.872901 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b" gracePeriod=600 Dec 01 19:50:14 crc kubenswrapper[4960]: I1201 19:50:14.540250 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b" exitCode=0 Dec 01 19:50:14 crc kubenswrapper[4960]: I1201 19:50:14.540317 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b"} Dec 01 19:50:14 crc kubenswrapper[4960]: I1201 19:50:14.541346 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d"} Dec 01 19:50:14 crc kubenswrapper[4960]: I1201 19:50:14.541419 4960 scope.go:117] "RemoveContainer" containerID="5ed3037f15601671b80ece4d091b7f5dc245d66c8ab67663d30f624d14052967" Dec 01 19:51:25 crc kubenswrapper[4960]: I1201 19:51:25.614484 4960 scope.go:117] "RemoveContainer" containerID="20b095bf7fda0ca7346692005f8329d69d71bf9677e4825beb65e368e1bf802f" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.355663 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k"] Dec 01 19:51:51 crc kubenswrapper[4960]: E1201 19:51:51.356499 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" containerName="registry" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.356519 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" containerName="registry" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.356644 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dd0037-d130-4fcf-8507-4dfa3059a9d9" containerName="registry" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.357619 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.361728 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.373732 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crf7j\" (UniqueName: \"kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.373798 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k"] Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.373817 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.373930 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.475102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crf7j\" (UniqueName: \"kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.475201 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.475285 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.476005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.476273 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.501029 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crf7j\" (UniqueName: \"kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.678187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:51 crc kubenswrapper[4960]: I1201 19:51:51.915259 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k"] Dec 01 19:51:52 crc kubenswrapper[4960]: I1201 19:51:52.332837 4960 generic.go:334] "Generic (PLEG): container finished" podID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerID="82023d1ce19f77ff1045907ea78a3e9f6dd280032bc9ff0bd575d25d3eaf05ef" exitCode=0 Dec 01 19:51:52 crc kubenswrapper[4960]: I1201 19:51:52.332919 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" event={"ID":"028186f6-ad65-47f2-be65-50d0e15dfc12","Type":"ContainerDied","Data":"82023d1ce19f77ff1045907ea78a3e9f6dd280032bc9ff0bd575d25d3eaf05ef"} Dec 01 19:51:52 crc kubenswrapper[4960]: I1201 19:51:52.333008 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" event={"ID":"028186f6-ad65-47f2-be65-50d0e15dfc12","Type":"ContainerStarted","Data":"a8379aa5ae6efacaef5b04572fdfdfb4b981df9ca95b0bfce8cc4a9c085727c7"} Dec 01 19:51:52 crc kubenswrapper[4960]: I1201 19:51:52.335517 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 19:51:54 crc kubenswrapper[4960]: I1201 19:51:54.348549 4960 generic.go:334] "Generic (PLEG): container finished" podID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerID="40020b2061215ca870d53db18936f414451e99ba1e3de6581cab2fa4979bead1" exitCode=0 Dec 01 19:51:54 crc kubenswrapper[4960]: I1201 19:51:54.348738 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" event={"ID":"028186f6-ad65-47f2-be65-50d0e15dfc12","Type":"ContainerDied","Data":"40020b2061215ca870d53db18936f414451e99ba1e3de6581cab2fa4979bead1"} Dec 01 19:51:55 crc kubenswrapper[4960]: I1201 19:51:55.361604 4960 generic.go:334] "Generic (PLEG): container finished" podID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerID="d17e91ff1d4a802f157bc45dc86d0f146998befcc97ccc681ce3cc4928116154" exitCode=0 Dec 01 19:51:55 crc kubenswrapper[4960]: I1201 19:51:55.361690 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" event={"ID":"028186f6-ad65-47f2-be65-50d0e15dfc12","Type":"ContainerDied","Data":"d17e91ff1d4a802f157bc45dc86d0f146998befcc97ccc681ce3cc4928116154"} Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.635098 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.750298 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crf7j\" (UniqueName: \"kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j\") pod \"028186f6-ad65-47f2-be65-50d0e15dfc12\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.750391 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle\") pod \"028186f6-ad65-47f2-be65-50d0e15dfc12\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.750456 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util\") pod \"028186f6-ad65-47f2-be65-50d0e15dfc12\" (UID: \"028186f6-ad65-47f2-be65-50d0e15dfc12\") " Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.755473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle" (OuterVolumeSpecName: "bundle") pod "028186f6-ad65-47f2-be65-50d0e15dfc12" (UID: "028186f6-ad65-47f2-be65-50d0e15dfc12"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.767834 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util" (OuterVolumeSpecName: "util") pod "028186f6-ad65-47f2-be65-50d0e15dfc12" (UID: "028186f6-ad65-47f2-be65-50d0e15dfc12"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.852070 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.852139 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/028186f6-ad65-47f2-be65-50d0e15dfc12-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:51:56 crc kubenswrapper[4960]: I1201 19:51:56.972867 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j" (OuterVolumeSpecName: "kube-api-access-crf7j") pod "028186f6-ad65-47f2-be65-50d0e15dfc12" (UID: "028186f6-ad65-47f2-be65-50d0e15dfc12"). InnerVolumeSpecName "kube-api-access-crf7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:51:57 crc kubenswrapper[4960]: I1201 19:51:57.055233 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crf7j\" (UniqueName: \"kubernetes.io/projected/028186f6-ad65-47f2-be65-50d0e15dfc12-kube-api-access-crf7j\") on node \"crc\" DevicePath \"\"" Dec 01 19:51:57 crc kubenswrapper[4960]: I1201 19:51:57.379845 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" event={"ID":"028186f6-ad65-47f2-be65-50d0e15dfc12","Type":"ContainerDied","Data":"a8379aa5ae6efacaef5b04572fdfdfb4b981df9ca95b0bfce8cc4a9c085727c7"} Dec 01 19:51:57 crc kubenswrapper[4960]: I1201 19:51:57.379931 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8379aa5ae6efacaef5b04572fdfdfb4b981df9ca95b0bfce8cc4a9c085727c7" Dec 01 19:51:57 crc kubenswrapper[4960]: I1201 19:51:57.380000 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.472704 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fdntq"] Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.473778 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-controller" containerID="cri-o://ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.473825 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.473926 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="northd" containerID="cri-o://eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.474050 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="sbdb" containerID="cri-o://498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.473997 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-acl-logging" containerID="cri-o://f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.474025 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-node" containerID="cri-o://427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.474198 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="nbdb" containerID="cri-o://a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.514473 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" containerID="cri-o://145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" gracePeriod=30 Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.802375 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/3.log" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.806207 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovn-acl-logging/0.log" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.807052 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovn-controller/0.log" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.807799 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869256 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqsbt"] Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869513 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869527 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869537 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kubecfg-setup" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869542 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kubecfg-setup" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869553 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869559 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869567 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869573 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869581 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="northd" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869587 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="northd" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869596 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="extract" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869602 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="extract" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869609 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-acl-logging" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869615 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-acl-logging" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869624 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="util" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869629 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="util" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869639 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="pull" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869644 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="pull" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869653 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869659 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869665 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869671 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869678 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="sbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869683 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="sbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869693 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869700 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869707 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="nbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869713 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="nbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869722 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-node" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869728 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-node" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869820 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869832 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869842 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="sbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869851 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-acl-logging" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869858 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="nbdb" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869865 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovn-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869874 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-ovn-metrics" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869882 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="kube-rbac-proxy-node" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869890 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" containerName="extract" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869898 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869906 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="northd" Dec 01 19:52:02 crc kubenswrapper[4960]: E1201 19:52:02.869986 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.869992 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.870094 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.870286 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerName="ovnkube-controller" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.871728 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937354 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937630 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937681 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rbhw\" (UniqueName: \"kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937715 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937745 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937779 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937834 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937857 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937924 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937953 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.937975 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938009 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938053 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938080 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938103 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938154 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938228 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938263 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.938309 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd\") pod \"8c9203b1-b24d-40f4-84e1-180293ad742b\" (UID: \"8c9203b1-b24d-40f4-84e1-180293ad742b\") " Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939605 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939659 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939687 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939716 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939742 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash" (OuterVolumeSpecName: "host-slash") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.939768 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket" (OuterVolumeSpecName: "log-socket") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.940473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.940523 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.940553 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.940577 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941009 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941054 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941081 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941690 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941715 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.941737 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log" (OuterVolumeSpecName: "node-log") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.942060 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.946526 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.947035 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw" (OuterVolumeSpecName: "kube-api-access-5rbhw") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "kube-api-access-5rbhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:52:02 crc kubenswrapper[4960]: I1201 19:52:02.957198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8c9203b1-b24d-40f4-84e1-180293ad742b" (UID: "8c9203b1-b24d-40f4-84e1-180293ad742b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039586 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-systemd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039648 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-kubelet\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039680 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-script-lib\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039705 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-var-lib-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039731 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-log-socket\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039752 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039784 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-netns\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039809 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-config\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039836 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-bin\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039867 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovn-node-metrics-cert\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039890 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-systemd-units\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039931 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b58ck\" (UniqueName: \"kubernetes.io/projected/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-kube-api-access-b58ck\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-etc-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.039986 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-ovn\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040009 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-env-overrides\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040028 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-slash\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040097 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-node-log\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040161 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040187 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-netd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040246 4960 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040261 4960 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040274 4960 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040287 4960 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040299 4960 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-node-log\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040311 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rbhw\" (UniqueName: \"kubernetes.io/projected/8c9203b1-b24d-40f4-84e1-180293ad742b-kube-api-access-5rbhw\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040323 4960 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040336 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040348 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c9203b1-b24d-40f4-84e1-180293ad742b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040359 4960 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040370 4960 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040382 4960 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040394 4960 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040405 4960 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-slash\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040417 4960 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-log-socket\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040428 4960 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040441 4960 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040454 4960 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040468 4960 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8c9203b1-b24d-40f4-84e1-180293ad742b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.040481 4960 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c9203b1-b24d-40f4-84e1-180293ad742b-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142585 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-etc-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142659 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-ovn\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142784 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-ovn\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142821 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-etc-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-env-overrides\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.142965 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-slash\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143040 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143109 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-node-log\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143141 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-slash\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143179 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143233 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-node-log\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143231 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143333 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-netd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143374 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-systemd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143288 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-kubelet\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143495 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-netd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143406 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-kubelet\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143478 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-systemd\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-script-lib\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-var-lib-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-log-socket\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143844 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143926 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-netns\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.143985 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-config\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-bin\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144156 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-env-overrides\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144197 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovn-node-metrics-cert\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144256 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-systemd-units\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144293 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-log-socket\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144302 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-var-lib-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144338 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-run-netns\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144358 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-run-openvswitch\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144402 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-systemd-units\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144380 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-host-cni-bin\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144315 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b58ck\" (UniqueName: \"kubernetes.io/projected/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-kube-api-access-b58ck\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.144653 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-script-lib\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.145560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovnkube-config\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.149346 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-ovn-node-metrics-cert\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.173495 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b58ck\" (UniqueName: \"kubernetes.io/projected/31b1f171-875f-4d1c-bdd1-6af1ce6bec29-kube-api-access-b58ck\") pod \"ovnkube-node-cqsbt\" (UID: \"31b1f171-875f-4d1c-bdd1-6af1ce6bec29\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.190202 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.420543 4960 generic.go:334] "Generic (PLEG): container finished" podID="31b1f171-875f-4d1c-bdd1-6af1ce6bec29" containerID="04ff843b70cecc9e2006137e9659e2b0086b82f4e6a6b769fe4302a5c8a5c0cb" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.420621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerDied","Data":"04ff843b70cecc9e2006137e9659e2b0086b82f4e6a6b769fe4302a5c8a5c0cb"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.421147 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"806a3a616ea458a454a59bb7421ea7995c58949de667dcac5d4f5a1356230b64"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.427284 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovnkube-controller/3.log" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.439550 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovn-acl-logging/0.log" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.442127 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-fdntq_8c9203b1-b24d-40f4-84e1-180293ad742b/ovn-controller/0.log" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445483 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445520 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445528 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445536 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445543 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445551 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" exitCode=0 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445559 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" exitCode=143 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445568 4960 generic.go:334] "Generic (PLEG): container finished" podID="8c9203b1-b24d-40f4-84e1-180293ad742b" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" exitCode=143 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445624 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445678 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445690 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445700 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445710 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445722 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445733 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445738 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445743 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445749 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445754 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445759 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445764 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445769 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445776 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445784 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445790 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445795 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445801 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445806 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445812 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445818 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445825 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445830 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445837 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445845 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445854 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445861 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445868 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445874 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445879 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445885 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445891 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445898 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445905 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445910 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" event={"ID":"8c9203b1-b24d-40f4-84e1-180293ad742b","Type":"ContainerDied","Data":"2cc943a224c9bac2a59edaefe4f1128741dd2f219f9e6194b54aec9ccb8e2c82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445930 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445937 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445942 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445948 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445953 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445958 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445964 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445970 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445975 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445980 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.445997 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.446063 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-fdntq" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.466906 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/2.log" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.467721 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/1.log" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.467776 4960 generic.go:334] "Generic (PLEG): container finished" podID="cfba83a1-281b-4f8f-be15-2760b2b9dfce" containerID="323e8982ac01bc4e25a3efd7a695c498c17fb5b84c9453497804840300393a6f" exitCode=2 Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.467816 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerDied","Data":"323e8982ac01bc4e25a3efd7a695c498c17fb5b84c9453497804840300393a6f"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.467845 4960 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b"} Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.468357 4960 scope.go:117] "RemoveContainer" containerID="323e8982ac01bc4e25a3efd7a695c498c17fb5b84c9453497804840300393a6f" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.491606 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.525268 4960 scope.go:117] "RemoveContainer" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.543862 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fdntq"] Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.550305 4960 scope.go:117] "RemoveContainer" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.557465 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-fdntq"] Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.594872 4960 scope.go:117] "RemoveContainer" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.647804 4960 scope.go:117] "RemoveContainer" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.686071 4960 scope.go:117] "RemoveContainer" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.713147 4960 scope.go:117] "RemoveContainer" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.728851 4960 scope.go:117] "RemoveContainer" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.745049 4960 scope.go:117] "RemoveContainer" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.778597 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.782625 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.782683 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} err="failed to get container status \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.782724 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.783337 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": container with ID starting with 3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb not found: ID does not exist" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.783390 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} err="failed to get container status \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": rpc error: code = NotFound desc = could not find container \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": container with ID starting with 3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.783432 4960 scope.go:117] "RemoveContainer" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.784054 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": container with ID starting with 498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3 not found: ID does not exist" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.784083 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} err="failed to get container status \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": rpc error: code = NotFound desc = could not find container \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": container with ID starting with 498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.784102 4960 scope.go:117] "RemoveContainer" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.785862 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": container with ID starting with a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82 not found: ID does not exist" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.785889 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} err="failed to get container status \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": rpc error: code = NotFound desc = could not find container \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": container with ID starting with a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.785906 4960 scope.go:117] "RemoveContainer" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.786269 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": container with ID starting with eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5 not found: ID does not exist" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786293 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} err="failed to get container status \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": rpc error: code = NotFound desc = could not find container \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": container with ID starting with eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786306 4960 scope.go:117] "RemoveContainer" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.786551 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": container with ID starting with d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426 not found: ID does not exist" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786574 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} err="failed to get container status \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": rpc error: code = NotFound desc = could not find container \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": container with ID starting with d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786588 4960 scope.go:117] "RemoveContainer" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.786926 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": container with ID starting with 427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e not found: ID does not exist" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786951 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} err="failed to get container status \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": rpc error: code = NotFound desc = could not find container \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": container with ID starting with 427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.786976 4960 scope.go:117] "RemoveContainer" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.787398 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": container with ID starting with f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757 not found: ID does not exist" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.787421 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} err="failed to get container status \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": rpc error: code = NotFound desc = could not find container \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": container with ID starting with f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.787434 4960 scope.go:117] "RemoveContainer" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.787686 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": container with ID starting with ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c not found: ID does not exist" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.787727 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} err="failed to get container status \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": rpc error: code = NotFound desc = could not find container \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": container with ID starting with ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.787744 4960 scope.go:117] "RemoveContainer" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: E1201 19:52:03.788082 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": container with ID starting with 21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2 not found: ID does not exist" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788122 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} err="failed to get container status \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": rpc error: code = NotFound desc = could not find container \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": container with ID starting with 21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788136 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788336 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} err="failed to get container status \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788361 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788578 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} err="failed to get container status \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": rpc error: code = NotFound desc = could not find container \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": container with ID starting with 3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788599 4960 scope.go:117] "RemoveContainer" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788791 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} err="failed to get container status \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": rpc error: code = NotFound desc = could not find container \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": container with ID starting with 498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.788812 4960 scope.go:117] "RemoveContainer" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789015 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} err="failed to get container status \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": rpc error: code = NotFound desc = could not find container \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": container with ID starting with a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789035 4960 scope.go:117] "RemoveContainer" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789276 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} err="failed to get container status \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": rpc error: code = NotFound desc = could not find container \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": container with ID starting with eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789296 4960 scope.go:117] "RemoveContainer" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789477 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} err="failed to get container status \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": rpc error: code = NotFound desc = could not find container \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": container with ID starting with d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.789496 4960 scope.go:117] "RemoveContainer" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.793290 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} err="failed to get container status \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": rpc error: code = NotFound desc = could not find container \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": container with ID starting with 427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.793342 4960 scope.go:117] "RemoveContainer" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.793826 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} err="failed to get container status \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": rpc error: code = NotFound desc = could not find container \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": container with ID starting with f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.793853 4960 scope.go:117] "RemoveContainer" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794051 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} err="failed to get container status \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": rpc error: code = NotFound desc = could not find container \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": container with ID starting with ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794070 4960 scope.go:117] "RemoveContainer" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794294 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} err="failed to get container status \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": rpc error: code = NotFound desc = could not find container \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": container with ID starting with 21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794314 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794512 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} err="failed to get container status \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794530 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794739 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} err="failed to get container status \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": rpc error: code = NotFound desc = could not find container \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": container with ID starting with 3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794759 4960 scope.go:117] "RemoveContainer" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794958 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} err="failed to get container status \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": rpc error: code = NotFound desc = could not find container \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": container with ID starting with 498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.794978 4960 scope.go:117] "RemoveContainer" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.799327 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} err="failed to get container status \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": rpc error: code = NotFound desc = could not find container \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": container with ID starting with a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.799393 4960 scope.go:117] "RemoveContainer" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.799841 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} err="failed to get container status \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": rpc error: code = NotFound desc = could not find container \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": container with ID starting with eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.799863 4960 scope.go:117] "RemoveContainer" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800210 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} err="failed to get container status \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": rpc error: code = NotFound desc = could not find container \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": container with ID starting with d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800234 4960 scope.go:117] "RemoveContainer" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800414 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} err="failed to get container status \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": rpc error: code = NotFound desc = could not find container \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": container with ID starting with 427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800434 4960 scope.go:117] "RemoveContainer" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800618 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} err="failed to get container status \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": rpc error: code = NotFound desc = could not find container \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": container with ID starting with f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800637 4960 scope.go:117] "RemoveContainer" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800807 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} err="failed to get container status \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": rpc error: code = NotFound desc = could not find container \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": container with ID starting with ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.800826 4960 scope.go:117] "RemoveContainer" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801026 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} err="failed to get container status \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": rpc error: code = NotFound desc = could not find container \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": container with ID starting with 21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801045 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801226 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} err="failed to get container status \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801244 4960 scope.go:117] "RemoveContainer" containerID="3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801403 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb"} err="failed to get container status \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": rpc error: code = NotFound desc = could not find container \"3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb\": container with ID starting with 3809f44e4ac652780ab2541f23c85a13da5322d26746cb5674e441ad52c1cddb not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801421 4960 scope.go:117] "RemoveContainer" containerID="498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801577 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3"} err="failed to get container status \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": rpc error: code = NotFound desc = could not find container \"498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3\": container with ID starting with 498b179b5ae50a1fe2377371a9f376b39edac976c48a3400bfdabc8e1daff2c3 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801596 4960 scope.go:117] "RemoveContainer" containerID="a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801775 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82"} err="failed to get container status \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": rpc error: code = NotFound desc = could not find container \"a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82\": container with ID starting with a047bec5b844a6f2791c12e3a36db737eba77a1d7d53be7847a771d535a5ba82 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.801794 4960 scope.go:117] "RemoveContainer" containerID="eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802002 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5"} err="failed to get container status \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": rpc error: code = NotFound desc = could not find container \"eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5\": container with ID starting with eb35452c1dd81b8591a556509ba63b11bc48592b0f68f1dfb38a2e5730c1e5b5 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802024 4960 scope.go:117] "RemoveContainer" containerID="d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802224 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426"} err="failed to get container status \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": rpc error: code = NotFound desc = could not find container \"d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426\": container with ID starting with d108736f060f7d42552e382c7d5fbf76c652dae3b4b68c39a332b4dd84ea6426 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802245 4960 scope.go:117] "RemoveContainer" containerID="427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802425 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e"} err="failed to get container status \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": rpc error: code = NotFound desc = could not find container \"427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e\": container with ID starting with 427026f946d268cffb9cf7e5aa1ee164c6c481ee92ac63fc2e077841c73ad30e not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802444 4960 scope.go:117] "RemoveContainer" containerID="f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802604 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757"} err="failed to get container status \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": rpc error: code = NotFound desc = could not find container \"f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757\": container with ID starting with f7beeec28216aaf7f10879b2169060a737a04460f2d226bffd24b5568e51b757 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802626 4960 scope.go:117] "RemoveContainer" containerID="ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802785 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c"} err="failed to get container status \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": rpc error: code = NotFound desc = could not find container \"ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c\": container with ID starting with ed1b27b0dea5ffc95a9096b070cfc4b85d15aee19fb1e9e80ea258fed39b448c not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802807 4960 scope.go:117] "RemoveContainer" containerID="21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802957 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2"} err="failed to get container status \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": rpc error: code = NotFound desc = could not find container \"21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2\": container with ID starting with 21b55d32ab982fbe6f4352a7fba6e4564b8916b10bbd271fb3cb0bf6255615e2 not found: ID does not exist" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.802976 4960 scope.go:117] "RemoveContainer" containerID="145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2" Dec 01 19:52:03 crc kubenswrapper[4960]: I1201 19:52:03.803146 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2"} err="failed to get container status \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": rpc error: code = NotFound desc = could not find container \"145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2\": container with ID starting with 145e84898618312bb69c8b3c19a5f71f02ca561356538495bce2c7241c10cae2 not found: ID does not exist" Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.477048 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/2.log" Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.478079 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/1.log" Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.478217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gns76" event={"ID":"cfba83a1-281b-4f8f-be15-2760b2b9dfce","Type":"ContainerStarted","Data":"ddc60c0d3b9da7ec467eb1583ad6d8e43f651fa3eb789bfffa1e6ff5c3419f13"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482488 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"06e548828b9bf8147e773d7a24fcab0364e325a8b3033b18b4b51807e9d20d1b"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482543 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"043f33b7433c753d6b7ef80810a9ac4b590e961ec6b7090dbee715857841c560"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482553 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"bb21072a4d66770cea7f585db43680e5f96294b8b976357f7e7416ad93b3cabd"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482564 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"35604f56b490c1c17ffd11ebb057ee7b0ddae3f13cdc1bca9d330cdba818c4bb"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482577 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"a20528f4cf14633119f3ddc09a814c82bfd4bcda7e5462f153f691b68e105a11"} Dec 01 19:52:04 crc kubenswrapper[4960]: I1201 19:52:04.482587 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"2ab450b7336ea538c59d8222ef14f0f0f8a9e88492d7d9054ad559d6c74626d4"} Dec 01 19:52:05 crc kubenswrapper[4960]: I1201 19:52:05.331750 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c9203b1-b24d-40f4-84e1-180293ad742b" path="/var/lib/kubelet/pods/8c9203b1-b24d-40f4-84e1-180293ad742b/volumes" Dec 01 19:52:05 crc kubenswrapper[4960]: I1201 19:52:05.617585 4960 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 01 19:52:07 crc kubenswrapper[4960]: I1201 19:52:07.504314 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"c5dfbcbafd9897f3aa0832903e75d4cef20e7b7ca2a10eb0a6a17af9d46e78df"} Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.391441 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf"] Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.392667 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.394361 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.394466 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-p5m8q" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.394889 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.427319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnsj8\" (UniqueName: \"kubernetes.io/projected/ba9ca50d-24cd-4d42-b130-25c0e744bcb4-kube-api-access-wnsj8\") pod \"obo-prometheus-operator-668cf9dfbb-dhxwf\" (UID: \"ba9ca50d-24cd-4d42-b130-25c0e744bcb4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.508326 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth"] Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.509078 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.510751 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.511614 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-b6dbm" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.517626 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg"] Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.518410 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.523241 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" event={"ID":"31b1f171-875f-4d1c-bdd1-6af1ce6bec29","Type":"ContainerStarted","Data":"dbacf43c42df284f44e1395e21f4b648185adc9081bbd5f0d4df6d0bbc71bad6"} Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.523563 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.523625 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.528493 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnsj8\" (UniqueName: \"kubernetes.io/projected/ba9ca50d-24cd-4d42-b130-25c0e744bcb4-kube-api-access-wnsj8\") pod \"obo-prometheus-operator-668cf9dfbb-dhxwf\" (UID: \"ba9ca50d-24cd-4d42-b130-25c0e744bcb4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.553906 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.576828 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnsj8\" (UniqueName: \"kubernetes.io/projected/ba9ca50d-24cd-4d42-b130-25c0e744bcb4-kube-api-access-wnsj8\") pod \"obo-prometheus-operator-668cf9dfbb-dhxwf\" (UID: \"ba9ca50d-24cd-4d42-b130-25c0e744bcb4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.604700 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" podStartSLOduration=7.604675784 podStartE2EDuration="7.604675784s" podCreationTimestamp="2025-12-01 19:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:52:09.604079955 +0000 UTC m=+764.891571634" watchObservedRunningTime="2025-12-01 19:52:09.604675784 +0000 UTC m=+764.892167453" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.629398 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.629507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.629575 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.629611 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.708570 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.730884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.730955 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.731005 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.731039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.735768 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.735853 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.736365 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6e2b5c4-73c4-45a8-8d81-08adbe28e39f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg\" (UID: \"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.736497 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/069dc362-519d-4349-b53e-288ee73b9916-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth\" (UID: \"069dc362-519d-4349-b53e-288ee73b9916\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.742258 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(868b054dcbb4c7c21c4fae8424a0229ebf30f8f20aa00fd21234c553ae140082): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.742343 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(868b054dcbb4c7c21c4fae8424a0229ebf30f8f20aa00fd21234c553ae140082): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.742368 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(868b054dcbb4c7c21c4fae8424a0229ebf30f8f20aa00fd21234c553ae140082): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.742417 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators(ba9ca50d-24cd-4d42-b130-25c0e744bcb4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators(ba9ca50d-24cd-4d42-b130-25c0e744bcb4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(868b054dcbb4c7c21c4fae8424a0229ebf30f8f20aa00fd21234c553ae140082): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" podUID="ba9ca50d-24cd-4d42-b130-25c0e744bcb4" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.762743 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-gfmgc"] Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.763470 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.780140 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-mxjx7" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.780293 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.826831 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.833056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.833140 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99x4j\" (UniqueName: \"kubernetes.io/projected/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-kube-api-access-99x4j\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.839723 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.866613 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(21df44c925c49872478914aa0eb59136ee6d2608b680bb7c7ebab2ee2e1e4c2c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.866728 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(21df44c925c49872478914aa0eb59136ee6d2608b680bb7c7ebab2ee2e1e4c2c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.866950 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(21df44c925c49872478914aa0eb59136ee6d2608b680bb7c7ebab2ee2e1e4c2c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.867012 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators(069dc362-519d-4349-b53e-288ee73b9916)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators(069dc362-519d-4349-b53e-288ee73b9916)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(21df44c925c49872478914aa0eb59136ee6d2608b680bb7c7ebab2ee2e1e4c2c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" podUID="069dc362-519d-4349-b53e-288ee73b9916" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.876899 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(85ea297a9bad32612590ac4af7076f81c6b09c5f3b34591e6d55780d52d68f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.876990 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(85ea297a9bad32612590ac4af7076f81c6b09c5f3b34591e6d55780d52d68f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.877018 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(85ea297a9bad32612590ac4af7076f81c6b09c5f3b34591e6d55780d52d68f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:09 crc kubenswrapper[4960]: E1201 19:52:09.877083 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators(d6e2b5c4-73c4-45a8-8d81-08adbe28e39f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators(d6e2b5c4-73c4-45a8-8d81-08adbe28e39f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(85ea297a9bad32612590ac4af7076f81c6b09c5f3b34591e6d55780d52d68f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" podUID="d6e2b5c4-73c4-45a8-8d81-08adbe28e39f" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.922540 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tktnr"] Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.923546 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.925932 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4sgv9" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.933887 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99x4j\" (UniqueName: \"kubernetes.io/projected/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-kube-api-access-99x4j\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.933963 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.937776 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:09 crc kubenswrapper[4960]: I1201 19:52:09.956744 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99x4j\" (UniqueName: \"kubernetes.io/projected/953224f9-3cdd-4e9c-98ac-aec3fc6ccc23-kube-api-access-99x4j\") pod \"observability-operator-d8bb48f5d-gfmgc\" (UID: \"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23\") " pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.035920 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f86337f-46f5-46aa-aaba-41179da85cab-openshift-service-ca\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.036104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6chg\" (UniqueName: \"kubernetes.io/projected/4f86337f-46f5-46aa-aaba-41179da85cab-kube-api-access-h6chg\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.082532 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.107436 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(d4b0f2aaf597d4a50e04d932c5224c20b6476b55666e1d4987db833e256b3e0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.107522 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(d4b0f2aaf597d4a50e04d932c5224c20b6476b55666e1d4987db833e256b3e0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.107553 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(d4b0f2aaf597d4a50e04d932c5224c20b6476b55666e1d4987db833e256b3e0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.107612 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-gfmgc_openshift-operators(953224f9-3cdd-4e9c-98ac-aec3fc6ccc23)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-gfmgc_openshift-operators(953224f9-3cdd-4e9c-98ac-aec3fc6ccc23)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(d4b0f2aaf597d4a50e04d932c5224c20b6476b55666e1d4987db833e256b3e0e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" podUID="953224f9-3cdd-4e9c-98ac-aec3fc6ccc23" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.136933 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6chg\" (UniqueName: \"kubernetes.io/projected/4f86337f-46f5-46aa-aaba-41179da85cab-kube-api-access-h6chg\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.137009 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f86337f-46f5-46aa-aaba-41179da85cab-openshift-service-ca\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.137917 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f86337f-46f5-46aa-aaba-41179da85cab-openshift-service-ca\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.157657 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6chg\" (UniqueName: \"kubernetes.io/projected/4f86337f-46f5-46aa-aaba-41179da85cab-kube-api-access-h6chg\") pod \"perses-operator-5446b9c989-tktnr\" (UID: \"4f86337f-46f5-46aa-aaba-41179da85cab\") " pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.239907 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.262214 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(b3b1d7e361fc25b394edbb23ed39f51648c200f59aa7eb83aaf9fb1c8748d6a5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.262317 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(b3b1d7e361fc25b394edbb23ed39f51648c200f59aa7eb83aaf9fb1c8748d6a5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.262352 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(b3b1d7e361fc25b394edbb23ed39f51648c200f59aa7eb83aaf9fb1c8748d6a5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.262442 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-tktnr_openshift-operators(4f86337f-46f5-46aa-aaba-41179da85cab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-tktnr_openshift-operators(4f86337f-46f5-46aa-aaba-41179da85cab)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(b3b1d7e361fc25b394edbb23ed39f51648c200f59aa7eb83aaf9fb1c8748d6a5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-tktnr" podUID="4f86337f-46f5-46aa-aaba-41179da85cab" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.528126 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.574642 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.923031 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf"] Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.923222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.923892 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.946295 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth"] Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.946445 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.946978 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.949796 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(8aa1cf2e3f1f3dffdb99837657aadfd041118b157e786363a69df8afa2cc8865): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.949848 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(8aa1cf2e3f1f3dffdb99837657aadfd041118b157e786363a69df8afa2cc8865): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.949871 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(8aa1cf2e3f1f3dffdb99837657aadfd041118b157e786363a69df8afa2cc8865): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.949914 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators(ba9ca50d-24cd-4d42-b130-25c0e744bcb4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators(ba9ca50d-24cd-4d42-b130-25c0e744bcb4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-dhxwf_openshift-operators_ba9ca50d-24cd-4d42-b130-25c0e744bcb4_0(8aa1cf2e3f1f3dffdb99837657aadfd041118b157e786363a69df8afa2cc8865): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" podUID="ba9ca50d-24cd-4d42-b130-25c0e744bcb4" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.974687 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(b24f73ce5e93ec1d363bd518980d39a4e615e5cdc58e2dfd671bb1b80a1fae88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.974766 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(b24f73ce5e93ec1d363bd518980d39a4e615e5cdc58e2dfd671bb1b80a1fae88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.974791 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(b24f73ce5e93ec1d363bd518980d39a4e615e5cdc58e2dfd671bb1b80a1fae88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:10 crc kubenswrapper[4960]: E1201 19:52:10.974836 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators(069dc362-519d-4349-b53e-288ee73b9916)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators(069dc362-519d-4349-b53e-288ee73b9916)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_openshift-operators_069dc362-519d-4349-b53e-288ee73b9916_0(b24f73ce5e93ec1d363bd518980d39a4e615e5cdc58e2dfd671bb1b80a1fae88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" podUID="069dc362-519d-4349-b53e-288ee73b9916" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.979083 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-gfmgc"] Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.979227 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.979688 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.997152 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg"] Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.997323 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:10 crc kubenswrapper[4960]: I1201 19:52:10.997759 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:11 crc kubenswrapper[4960]: I1201 19:52:11.006780 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tktnr"] Dec 01 19:52:11 crc kubenswrapper[4960]: I1201 19:52:11.006914 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:11 crc kubenswrapper[4960]: I1201 19:52:11.007373 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.029246 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(e879c3ced1d09d20e8aa35e9f32ba75843340aee950790b7e67a4bbc51306f85): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.030966 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(e879c3ced1d09d20e8aa35e9f32ba75843340aee950790b7e67a4bbc51306f85): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.031001 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(e879c3ced1d09d20e8aa35e9f32ba75843340aee950790b7e67a4bbc51306f85): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.031060 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-gfmgc_openshift-operators(953224f9-3cdd-4e9c-98ac-aec3fc6ccc23)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-gfmgc_openshift-operators(953224f9-3cdd-4e9c-98ac-aec3fc6ccc23)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-gfmgc_openshift-operators_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23_0(e879c3ced1d09d20e8aa35e9f32ba75843340aee950790b7e67a4bbc51306f85): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" podUID="953224f9-3cdd-4e9c-98ac-aec3fc6ccc23" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.075300 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(3d161c4b0f5f20043956cf08ae0121553186d7d857f8f61ceb16480b497c05ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.075377 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(3d161c4b0f5f20043956cf08ae0121553186d7d857f8f61ceb16480b497c05ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.075403 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(3d161c4b0f5f20043956cf08ae0121553186d7d857f8f61ceb16480b497c05ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.075457 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators(d6e2b5c4-73c4-45a8-8d81-08adbe28e39f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators(d6e2b5c4-73c4-45a8-8d81-08adbe28e39f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_openshift-operators_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f_0(3d161c4b0f5f20043956cf08ae0121553186d7d857f8f61ceb16480b497c05ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" podUID="d6e2b5c4-73c4-45a8-8d81-08adbe28e39f" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.089221 4960 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(53025a32b6ffaae773d48b79eeddebfad78c4d89f7c1dad480bf6088da2c28f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.089296 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(53025a32b6ffaae773d48b79eeddebfad78c4d89f7c1dad480bf6088da2c28f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.089324 4960 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(53025a32b6ffaae773d48b79eeddebfad78c4d89f7c1dad480bf6088da2c28f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:11 crc kubenswrapper[4960]: E1201 19:52:11.089372 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-tktnr_openshift-operators(4f86337f-46f5-46aa-aaba-41179da85cab)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-tktnr_openshift-operators(4f86337f-46f5-46aa-aaba-41179da85cab)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-tktnr_openshift-operators_4f86337f-46f5-46aa-aaba-41179da85cab_0(53025a32b6ffaae773d48b79eeddebfad78c4d89f7c1dad480bf6088da2c28f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-tktnr" podUID="4f86337f-46f5-46aa-aaba-41179da85cab" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.323726 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.323813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.323813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.325014 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.325297 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.325333 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.676269 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf"] Dec 01 19:52:21 crc kubenswrapper[4960]: W1201 19:52:21.682012 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba9ca50d_24cd_4d42_b130_25c0e744bcb4.slice/crio-e0f9dbd1f14595e72381ca12b62a699edb80e7fd6ef7095419c20c435a461bf4 WatchSource:0}: Error finding container e0f9dbd1f14595e72381ca12b62a699edb80e7fd6ef7095419c20c435a461bf4: Status 404 returned error can't find the container with id e0f9dbd1f14595e72381ca12b62a699edb80e7fd6ef7095419c20c435a461bf4 Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.689746 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg"] Dec 01 19:52:21 crc kubenswrapper[4960]: W1201 19:52:21.702347 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6e2b5c4_73c4_45a8_8d81_08adbe28e39f.slice/crio-0dee39da11468e19b874e5d52051d88f3e00554ffc67245e43066fc20290a125 WatchSource:0}: Error finding container 0dee39da11468e19b874e5d52051d88f3e00554ffc67245e43066fc20290a125: Status 404 returned error can't find the container with id 0dee39da11468e19b874e5d52051d88f3e00554ffc67245e43066fc20290a125 Dec 01 19:52:21 crc kubenswrapper[4960]: I1201 19:52:21.805640 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth"] Dec 01 19:52:21 crc kubenswrapper[4960]: W1201 19:52:21.812419 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod069dc362_519d_4349_b53e_288ee73b9916.slice/crio-31fd7c8dfdda050bce69697d717c50fe0918c9e34fe1e4ce733e792346b33602 WatchSource:0}: Error finding container 31fd7c8dfdda050bce69697d717c50fe0918c9e34fe1e4ce733e792346b33602: Status 404 returned error can't find the container with id 31fd7c8dfdda050bce69697d717c50fe0918c9e34fe1e4ce733e792346b33602 Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.324320 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.324926 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.553816 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-tktnr"] Dec 01 19:52:22 crc kubenswrapper[4960]: W1201 19:52:22.559730 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f86337f_46f5_46aa_aaba_41179da85cab.slice/crio-abc52797501c202f67cfc06ca0120e5fe5b5154267e510463a4107a3d3202905 WatchSource:0}: Error finding container abc52797501c202f67cfc06ca0120e5fe5b5154267e510463a4107a3d3202905: Status 404 returned error can't find the container with id abc52797501c202f67cfc06ca0120e5fe5b5154267e510463a4107a3d3202905 Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.605082 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" event={"ID":"069dc362-519d-4349-b53e-288ee73b9916","Type":"ContainerStarted","Data":"31fd7c8dfdda050bce69697d717c50fe0918c9e34fe1e4ce733e792346b33602"} Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.606616 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" event={"ID":"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f","Type":"ContainerStarted","Data":"0dee39da11468e19b874e5d52051d88f3e00554ffc67245e43066fc20290a125"} Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.608251 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tktnr" event={"ID":"4f86337f-46f5-46aa-aaba-41179da85cab","Type":"ContainerStarted","Data":"abc52797501c202f67cfc06ca0120e5fe5b5154267e510463a4107a3d3202905"} Dec 01 19:52:22 crc kubenswrapper[4960]: I1201 19:52:22.609609 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" event={"ID":"ba9ca50d-24cd-4d42-b130-25c0e744bcb4","Type":"ContainerStarted","Data":"e0f9dbd1f14595e72381ca12b62a699edb80e7fd6ef7095419c20c435a461bf4"} Dec 01 19:52:25 crc kubenswrapper[4960]: I1201 19:52:25.665606 4960 scope.go:117] "RemoveContainer" containerID="9893949042e7a699aeb007b731e9d5645ed67157dafa719135f7f12f138b6b7b" Dec 01 19:52:26 crc kubenswrapper[4960]: I1201 19:52:26.323623 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:26 crc kubenswrapper[4960]: I1201 19:52:26.324439 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:26 crc kubenswrapper[4960]: I1201 19:52:26.638957 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gns76_cfba83a1-281b-4f8f-be15-2760b2b9dfce/kube-multus/2.log" Dec 01 19:52:31 crc kubenswrapper[4960]: I1201 19:52:31.740590 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-gfmgc"] Dec 01 19:52:31 crc kubenswrapper[4960]: W1201 19:52:31.752247 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod953224f9_3cdd_4e9c_98ac_aec3fc6ccc23.slice/crio-b65ebbd78e66a6dae37b35f197141cfa76f499851e63b4186fe033e7fe157014 WatchSource:0}: Error finding container b65ebbd78e66a6dae37b35f197141cfa76f499851e63b4186fe033e7fe157014: Status 404 returned error can't find the container with id b65ebbd78e66a6dae37b35f197141cfa76f499851e63b4186fe033e7fe157014 Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.679445 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" event={"ID":"ba9ca50d-24cd-4d42-b130-25c0e744bcb4","Type":"ContainerStarted","Data":"c3e7ebf924ef1f77b22b773efc8c8fd3f794cf49a0cc76160c937ec747737ac2"} Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.682699 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" event={"ID":"069dc362-519d-4349-b53e-288ee73b9916","Type":"ContainerStarted","Data":"500c2834612f22a5d4dc78a328f0944f19e59d60fb546db8e0b6787e2c2e78c9"} Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.685099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" event={"ID":"d6e2b5c4-73c4-45a8-8d81-08adbe28e39f","Type":"ContainerStarted","Data":"225be4716ac45d455eed5162dd13888aafc1734bde434736602dab6fcb8ce5ce"} Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.687263 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-tktnr" event={"ID":"4f86337f-46f5-46aa-aaba-41179da85cab","Type":"ContainerStarted","Data":"c2c29f75a37dab822914709f6a9260e1113bac9ad405855ede08553e47a3529b"} Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.688022 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.689550 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" event={"ID":"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23","Type":"ContainerStarted","Data":"b65ebbd78e66a6dae37b35f197141cfa76f499851e63b4186fe033e7fe157014"} Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.703499 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-dhxwf" podStartSLOduration=13.848028853 podStartE2EDuration="23.70346697s" podCreationTimestamp="2025-12-01 19:52:09 +0000 UTC" firstStartedPulling="2025-12-01 19:52:21.686365045 +0000 UTC m=+776.973856714" lastFinishedPulling="2025-12-01 19:52:31.541803162 +0000 UTC m=+786.829294831" observedRunningTime="2025-12-01 19:52:32.698844917 +0000 UTC m=+787.986336596" watchObservedRunningTime="2025-12-01 19:52:32.70346697 +0000 UTC m=+787.990958639" Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.724390 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-smxth" podStartSLOduration=14.042794255 podStartE2EDuration="23.72436869s" podCreationTimestamp="2025-12-01 19:52:09 +0000 UTC" firstStartedPulling="2025-12-01 19:52:21.816533501 +0000 UTC m=+777.104025170" lastFinishedPulling="2025-12-01 19:52:31.498107926 +0000 UTC m=+786.785599605" observedRunningTime="2025-12-01 19:52:32.722585564 +0000 UTC m=+788.010077253" watchObservedRunningTime="2025-12-01 19:52:32.72436869 +0000 UTC m=+788.011860359" Dec 01 19:52:32 crc kubenswrapper[4960]: I1201 19:52:32.746334 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg" podStartSLOduration=13.909814566 podStartE2EDuration="23.746307211s" podCreationTimestamp="2025-12-01 19:52:09 +0000 UTC" firstStartedPulling="2025-12-01 19:52:21.705286907 +0000 UTC m=+776.992778576" lastFinishedPulling="2025-12-01 19:52:31.541779552 +0000 UTC m=+786.829271221" observedRunningTime="2025-12-01 19:52:32.745841386 +0000 UTC m=+788.033333055" watchObservedRunningTime="2025-12-01 19:52:32.746307211 +0000 UTC m=+788.033798880" Dec 01 19:52:33 crc kubenswrapper[4960]: I1201 19:52:33.247013 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqsbt" Dec 01 19:52:33 crc kubenswrapper[4960]: I1201 19:52:33.298084 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-tktnr" podStartSLOduration=15.318386184 podStartE2EDuration="24.298029461s" podCreationTimestamp="2025-12-01 19:52:09 +0000 UTC" firstStartedPulling="2025-12-01 19:52:22.563126116 +0000 UTC m=+777.850617785" lastFinishedPulling="2025-12-01 19:52:31.542769403 +0000 UTC m=+786.830261062" observedRunningTime="2025-12-01 19:52:32.766823438 +0000 UTC m=+788.054315107" watchObservedRunningTime="2025-12-01 19:52:33.298029461 +0000 UTC m=+788.585521130" Dec 01 19:52:37 crc kubenswrapper[4960]: I1201 19:52:37.723768 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" event={"ID":"953224f9-3cdd-4e9c-98ac-aec3fc6ccc23","Type":"ContainerStarted","Data":"535f1cbedac730a850ce534d6681029382c0cb2b7814c304acd9f48657dc000b"} Dec 01 19:52:37 crc kubenswrapper[4960]: I1201 19:52:37.724533 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:37 crc kubenswrapper[4960]: I1201 19:52:37.728097 4960 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-gfmgc container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.12:8081/healthz\": dial tcp 10.217.0.12:8081: connect: connection refused" start-of-body= Dec 01 19:52:37 crc kubenswrapper[4960]: I1201 19:52:37.728238 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" podUID="953224f9-3cdd-4e9c-98ac-aec3fc6ccc23" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.12:8081/healthz\": dial tcp 10.217.0.12:8081: connect: connection refused" Dec 01 19:52:37 crc kubenswrapper[4960]: I1201 19:52:37.763360 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" podStartSLOduration=23.173413344 podStartE2EDuration="28.763338222s" podCreationTimestamp="2025-12-01 19:52:09 +0000 UTC" firstStartedPulling="2025-12-01 19:52:31.756485679 +0000 UTC m=+787.043977338" lastFinishedPulling="2025-12-01 19:52:37.346410537 +0000 UTC m=+792.633902216" observedRunningTime="2025-12-01 19:52:37.761531676 +0000 UTC m=+793.049023375" watchObservedRunningTime="2025-12-01 19:52:37.763338222 +0000 UTC m=+793.050829901" Dec 01 19:52:38 crc kubenswrapper[4960]: I1201 19:52:38.779863 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-gfmgc" Dec 01 19:52:40 crc kubenswrapper[4960]: I1201 19:52:40.245555 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-tktnr" Dec 01 19:52:43 crc kubenswrapper[4960]: I1201 19:52:43.871195 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:52:43 crc kubenswrapper[4960]: I1201 19:52:43.872363 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.789472 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-9lvxq"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.790224 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.793243 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.793393 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.796623 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-htnmd" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.803831 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-9lvxq"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.814009 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7dpzk"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.815087 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7dpzk" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.818427 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-b7fth" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.820084 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsw9v"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.820971 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.824518 4960 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7dhz8" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.827659 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7dpzk"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.864696 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsw9v"] Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.881960 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2z8f\" (UniqueName: \"kubernetes.io/projected/2e0d8acc-f304-45d4-b53c-dec0723c94e2-kube-api-access-x2z8f\") pod \"cert-manager-5b446d88c5-7dpzk\" (UID: \"2e0d8acc-f304-45d4-b53c-dec0723c94e2\") " pod="cert-manager/cert-manager-5b446d88c5-7dpzk" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.882035 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j66m9\" (UniqueName: \"kubernetes.io/projected/12db57f7-a5f5-4a77-a412-d659e1490968-kube-api-access-j66m9\") pod \"cert-manager-webhook-5655c58dd6-fsw9v\" (UID: \"12db57f7-a5f5-4a77-a412-d659e1490968\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.882069 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xkxg\" (UniqueName: \"kubernetes.io/projected/75151dc8-631d-42be-9b8c-e6b9a532c0da-kube-api-access-4xkxg\") pod \"cert-manager-cainjector-7f985d654d-9lvxq\" (UID: \"75151dc8-631d-42be-9b8c-e6b9a532c0da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.983182 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2z8f\" (UniqueName: \"kubernetes.io/projected/2e0d8acc-f304-45d4-b53c-dec0723c94e2-kube-api-access-x2z8f\") pod \"cert-manager-5b446d88c5-7dpzk\" (UID: \"2e0d8acc-f304-45d4-b53c-dec0723c94e2\") " pod="cert-manager/cert-manager-5b446d88c5-7dpzk" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.983246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j66m9\" (UniqueName: \"kubernetes.io/projected/12db57f7-a5f5-4a77-a412-d659e1490968-kube-api-access-j66m9\") pod \"cert-manager-webhook-5655c58dd6-fsw9v\" (UID: \"12db57f7-a5f5-4a77-a412-d659e1490968\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:45 crc kubenswrapper[4960]: I1201 19:52:45.983736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xkxg\" (UniqueName: \"kubernetes.io/projected/75151dc8-631d-42be-9b8c-e6b9a532c0da-kube-api-access-4xkxg\") pod \"cert-manager-cainjector-7f985d654d-9lvxq\" (UID: \"75151dc8-631d-42be-9b8c-e6b9a532c0da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.005087 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xkxg\" (UniqueName: \"kubernetes.io/projected/75151dc8-631d-42be-9b8c-e6b9a532c0da-kube-api-access-4xkxg\") pod \"cert-manager-cainjector-7f985d654d-9lvxq\" (UID: \"75151dc8-631d-42be-9b8c-e6b9a532c0da\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.005187 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j66m9\" (UniqueName: \"kubernetes.io/projected/12db57f7-a5f5-4a77-a412-d659e1490968-kube-api-access-j66m9\") pod \"cert-manager-webhook-5655c58dd6-fsw9v\" (UID: \"12db57f7-a5f5-4a77-a412-d659e1490968\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.013749 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2z8f\" (UniqueName: \"kubernetes.io/projected/2e0d8acc-f304-45d4-b53c-dec0723c94e2-kube-api-access-x2z8f\") pod \"cert-manager-5b446d88c5-7dpzk\" (UID: \"2e0d8acc-f304-45d4-b53c-dec0723c94e2\") " pod="cert-manager/cert-manager-5b446d88c5-7dpzk" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.110982 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.134463 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7dpzk" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.147276 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.363262 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-9lvxq"] Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.441764 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7dpzk"] Dec 01 19:52:46 crc kubenswrapper[4960]: W1201 19:52:46.446095 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e0d8acc_f304_45d4_b53c_dec0723c94e2.slice/crio-2487511f1d9ee67a846dbf00ce5b3683671e9c455fc3e7f8f7b7b6ad18ec9432 WatchSource:0}: Error finding container 2487511f1d9ee67a846dbf00ce5b3683671e9c455fc3e7f8f7b7b6ad18ec9432: Status 404 returned error can't find the container with id 2487511f1d9ee67a846dbf00ce5b3683671e9c455fc3e7f8f7b7b6ad18ec9432 Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.461300 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-fsw9v"] Dec 01 19:52:46 crc kubenswrapper[4960]: W1201 19:52:46.469048 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12db57f7_a5f5_4a77_a412_d659e1490968.slice/crio-e2973e06f98a52ab4d15bb76ef04c0d4d9700293a0238b34aec370bb0a1cebe7 WatchSource:0}: Error finding container e2973e06f98a52ab4d15bb76ef04c0d4d9700293a0238b34aec370bb0a1cebe7: Status 404 returned error can't find the container with id e2973e06f98a52ab4d15bb76ef04c0d4d9700293a0238b34aec370bb0a1cebe7 Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.782262 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" event={"ID":"12db57f7-a5f5-4a77-a412-d659e1490968","Type":"ContainerStarted","Data":"e2973e06f98a52ab4d15bb76ef04c0d4d9700293a0238b34aec370bb0a1cebe7"} Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.787264 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7dpzk" event={"ID":"2e0d8acc-f304-45d4-b53c-dec0723c94e2","Type":"ContainerStarted","Data":"2487511f1d9ee67a846dbf00ce5b3683671e9c455fc3e7f8f7b7b6ad18ec9432"} Dec 01 19:52:46 crc kubenswrapper[4960]: I1201 19:52:46.788975 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" event={"ID":"75151dc8-631d-42be-9b8c-e6b9a532c0da","Type":"ContainerStarted","Data":"2fdb51465e3e6dc61d0022a10d19307fb26413debd6edce26244195caa5ac0d5"} Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.822759 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" event={"ID":"12db57f7-a5f5-4a77-a412-d659e1490968","Type":"ContainerStarted","Data":"7aab87c84e85c4f888aa05a199b72657183b1e996bb2bd7b289047890fad8044"} Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.823801 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.825594 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7dpzk" event={"ID":"2e0d8acc-f304-45d4-b53c-dec0723c94e2","Type":"ContainerStarted","Data":"e57a48a28f22a6579c5085c5e2fa79bd139b291e0658db5d721b6b9919f90c24"} Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.829471 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" event={"ID":"75151dc8-631d-42be-9b8c-e6b9a532c0da","Type":"ContainerStarted","Data":"d3b1d37d89f4b51d60c7f858deb974746abea0bd00e839d4ae949e8689348128"} Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.843083 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" podStartSLOduration=2.234292083 podStartE2EDuration="5.843054879s" podCreationTimestamp="2025-12-01 19:52:45 +0000 UTC" firstStartedPulling="2025-12-01 19:52:46.471393965 +0000 UTC m=+801.758885634" lastFinishedPulling="2025-12-01 19:52:50.080156761 +0000 UTC m=+805.367648430" observedRunningTime="2025-12-01 19:52:50.837361052 +0000 UTC m=+806.124852751" watchObservedRunningTime="2025-12-01 19:52:50.843054879 +0000 UTC m=+806.130546588" Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.859759 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-9lvxq" podStartSLOduration=2.222589108 podStartE2EDuration="5.859726636s" podCreationTimestamp="2025-12-01 19:52:45 +0000 UTC" firstStartedPulling="2025-12-01 19:52:46.37815615 +0000 UTC m=+801.665647819" lastFinishedPulling="2025-12-01 19:52:50.015293668 +0000 UTC m=+805.302785347" observedRunningTime="2025-12-01 19:52:50.855530905 +0000 UTC m=+806.143022584" watchObservedRunningTime="2025-12-01 19:52:50.859726636 +0000 UTC m=+806.147218335" Dec 01 19:52:50 crc kubenswrapper[4960]: I1201 19:52:50.878037 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7dpzk" podStartSLOduration=2.30284429 podStartE2EDuration="5.878015074s" podCreationTimestamp="2025-12-01 19:52:45 +0000 UTC" firstStartedPulling="2025-12-01 19:52:46.447627637 +0000 UTC m=+801.735119296" lastFinishedPulling="2025-12-01 19:52:50.022798411 +0000 UTC m=+805.310290080" observedRunningTime="2025-12-01 19:52:50.87563338 +0000 UTC m=+806.163125049" watchObservedRunningTime="2025-12-01 19:52:50.878015074 +0000 UTC m=+806.165506743" Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.800533 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.802293 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.812601 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.898974 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2562v\" (UniqueName: \"kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.899282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:52 crc kubenswrapper[4960]: I1201 19:52:52.899376 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.000866 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.000958 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.001013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2562v\" (UniqueName: \"kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.001454 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.001649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.022468 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2562v\" (UniqueName: \"kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v\") pod \"redhat-marketplace-kk8nr\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.121542 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.356156 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.852355 4960 generic.go:334] "Generic (PLEG): container finished" podID="8945da28-cded-473c-a097-c860c0f3a369" containerID="69f8fcd866b923ddec9bd2119e37490b39a8e8b4225b6d3f235d7c3e338de6fb" exitCode=0 Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.852412 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerDied","Data":"69f8fcd866b923ddec9bd2119e37490b39a8e8b4225b6d3f235d7c3e338de6fb"} Dec 01 19:52:53 crc kubenswrapper[4960]: I1201 19:52:53.852442 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerStarted","Data":"2a02888adc9b48669f1a247f259bdd97acd1b2431a8987be5e1492dd4a316950"} Dec 01 19:52:56 crc kubenswrapper[4960]: I1201 19:52:56.032493 4960 generic.go:334] "Generic (PLEG): container finished" podID="8945da28-cded-473c-a097-c860c0f3a369" containerID="9fa9c45ac36f7241f04d5b0a94839d90b9187fc1a106b0c81188e373277753d6" exitCode=0 Dec 01 19:52:56 crc kubenswrapper[4960]: I1201 19:52:56.032609 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerDied","Data":"9fa9c45ac36f7241f04d5b0a94839d90b9187fc1a106b0c81188e373277753d6"} Dec 01 19:52:56 crc kubenswrapper[4960]: I1201 19:52:56.151464 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-fsw9v" Dec 01 19:52:57 crc kubenswrapper[4960]: I1201 19:52:57.041001 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerStarted","Data":"e4c5c6766e26c4cb20904315f60ddc659adf1309862c5a8819b03147ba31db1c"} Dec 01 19:52:57 crc kubenswrapper[4960]: I1201 19:52:57.062229 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kk8nr" podStartSLOduration=2.458409749 podStartE2EDuration="5.062202733s" podCreationTimestamp="2025-12-01 19:52:52 +0000 UTC" firstStartedPulling="2025-12-01 19:52:53.855142139 +0000 UTC m=+809.142633828" lastFinishedPulling="2025-12-01 19:52:56.458935143 +0000 UTC m=+811.746426812" observedRunningTime="2025-12-01 19:52:57.061841802 +0000 UTC m=+812.349333511" watchObservedRunningTime="2025-12-01 19:52:57.062202733 +0000 UTC m=+812.349694422" Dec 01 19:53:03 crc kubenswrapper[4960]: I1201 19:53:03.122228 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:03 crc kubenswrapper[4960]: I1201 19:53:03.122303 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:03 crc kubenswrapper[4960]: I1201 19:53:03.202991 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:04 crc kubenswrapper[4960]: I1201 19:53:04.140222 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:05 crc kubenswrapper[4960]: I1201 19:53:05.585337 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:53:06 crc kubenswrapper[4960]: I1201 19:53:06.108273 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kk8nr" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="registry-server" containerID="cri-o://e4c5c6766e26c4cb20904315f60ddc659adf1309862c5a8819b03147ba31db1c" gracePeriod=2 Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.126693 4960 generic.go:334] "Generic (PLEG): container finished" podID="8945da28-cded-473c-a097-c860c0f3a369" containerID="e4c5c6766e26c4cb20904315f60ddc659adf1309862c5a8819b03147ba31db1c" exitCode=0 Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.126821 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerDied","Data":"e4c5c6766e26c4cb20904315f60ddc659adf1309862c5a8819b03147ba31db1c"} Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.395573 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.498683 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content\") pod \"8945da28-cded-473c-a097-c860c0f3a369\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.498778 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities\") pod \"8945da28-cded-473c-a097-c860c0f3a369\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.498841 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2562v\" (UniqueName: \"kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v\") pod \"8945da28-cded-473c-a097-c860c0f3a369\" (UID: \"8945da28-cded-473c-a097-c860c0f3a369\") " Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.500945 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities" (OuterVolumeSpecName: "utilities") pod "8945da28-cded-473c-a097-c860c0f3a369" (UID: "8945da28-cded-473c-a097-c860c0f3a369"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.508300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v" (OuterVolumeSpecName: "kube-api-access-2562v") pod "8945da28-cded-473c-a097-c860c0f3a369" (UID: "8945da28-cded-473c-a097-c860c0f3a369"). InnerVolumeSpecName "kube-api-access-2562v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.525504 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8945da28-cded-473c-a097-c860c0f3a369" (UID: "8945da28-cded-473c-a097-c860c0f3a369"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.602520 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.602592 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2562v\" (UniqueName: \"kubernetes.io/projected/8945da28-cded-473c-a097-c860c0f3a369-kube-api-access-2562v\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:08 crc kubenswrapper[4960]: I1201 19:53:08.602612 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8945da28-cded-473c-a097-c860c0f3a369-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.138202 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kk8nr" event={"ID":"8945da28-cded-473c-a097-c860c0f3a369","Type":"ContainerDied","Data":"2a02888adc9b48669f1a247f259bdd97acd1b2431a8987be5e1492dd4a316950"} Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.138289 4960 scope.go:117] "RemoveContainer" containerID="e4c5c6766e26c4cb20904315f60ddc659adf1309862c5a8819b03147ba31db1c" Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.138374 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kk8nr" Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.170188 4960 scope.go:117] "RemoveContainer" containerID="9fa9c45ac36f7241f04d5b0a94839d90b9187fc1a106b0c81188e373277753d6" Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.192296 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.195494 4960 scope.go:117] "RemoveContainer" containerID="69f8fcd866b923ddec9bd2119e37490b39a8e8b4225b6d3f235d7c3e338de6fb" Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.197836 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kk8nr"] Dec 01 19:53:09 crc kubenswrapper[4960]: I1201 19:53:09.330049 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8945da28-cded-473c-a097-c860c0f3a369" path="/var/lib/kubelet/pods/8945da28-cded-473c-a097-c860c0f3a369/volumes" Dec 01 19:53:13 crc kubenswrapper[4960]: I1201 19:53:13.870813 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:53:13 crc kubenswrapper[4960]: I1201 19:53:13.870912 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.071386 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf"] Dec 01 19:53:26 crc kubenswrapper[4960]: E1201 19:53:26.072169 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="registry-server" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.072191 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="registry-server" Dec 01 19:53:26 crc kubenswrapper[4960]: E1201 19:53:26.072233 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="extract-utilities" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.072247 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="extract-utilities" Dec 01 19:53:26 crc kubenswrapper[4960]: E1201 19:53:26.072270 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="extract-content" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.072284 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="extract-content" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.072456 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8945da28-cded-473c-a097-c860c0f3a369" containerName="registry-server" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.073987 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.076606 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.083478 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf"] Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.145135 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7blr\" (UniqueName: \"kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.145203 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.145484 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.246773 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7blr\" (UniqueName: \"kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.246854 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.246988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.247605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.247644 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.283057 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7blr\" (UniqueName: \"kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr\") pod \"142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.394984 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:26 crc kubenswrapper[4960]: I1201 19:53:26.658337 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf"] Dec 01 19:53:26 crc kubenswrapper[4960]: W1201 19:53:26.665189 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f4799aa_209e_459d_a9e4_c2221c6a7e33.slice/crio-a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4 WatchSource:0}: Error finding container a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4: Status 404 returned error can't find the container with id a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4 Dec 01 19:53:27 crc kubenswrapper[4960]: I1201 19:53:27.265481 4960 generic.go:334] "Generic (PLEG): container finished" podID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerID="463654f6c57cac8b5cf868a0c7dd34808c1c6f63b99ad29e3848167b1a08b0e8" exitCode=0 Dec 01 19:53:27 crc kubenswrapper[4960]: I1201 19:53:27.265541 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" event={"ID":"0f4799aa-209e-459d-a9e4-c2221c6a7e33","Type":"ContainerDied","Data":"463654f6c57cac8b5cf868a0c7dd34808c1c6f63b99ad29e3848167b1a08b0e8"} Dec 01 19:53:27 crc kubenswrapper[4960]: I1201 19:53:27.265610 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" event={"ID":"0f4799aa-209e-459d-a9e4-c2221c6a7e33","Type":"ContainerStarted","Data":"a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4"} Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.174182 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.175890 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.178175 4960 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-5gh7n" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.178532 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.179418 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.184477 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.380810 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4w6p\" (UniqueName: \"kubernetes.io/projected/fcd5c18b-c010-4158-9935-033a7f0e95a2-kube-api-access-l4w6p\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.380998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-748026ca-e157-4ebb-8901-c019b4b83025\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-748026ca-e157-4ebb-8901-c019b4b83025\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.482733 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-748026ca-e157-4ebb-8901-c019b4b83025\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-748026ca-e157-4ebb-8901-c019b4b83025\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.482883 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4w6p\" (UniqueName: \"kubernetes.io/projected/fcd5c18b-c010-4158-9935-033a7f0e95a2-kube-api-access-l4w6p\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.487329 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.487392 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-748026ca-e157-4ebb-8901-c019b4b83025\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-748026ca-e157-4ebb-8901-c019b4b83025\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3f6a0d787937d059b59573109fa50a12bdc180cd58b79dbc2db209ec1a134c52/globalmount\"" pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.505281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4w6p\" (UniqueName: \"kubernetes.io/projected/fcd5c18b-c010-4158-9935-033a7f0e95a2-kube-api-access-l4w6p\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.518277 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-748026ca-e157-4ebb-8901-c019b4b83025\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-748026ca-e157-4ebb-8901-c019b4b83025\") pod \"minio\" (UID: \"fcd5c18b-c010-4158-9935-033a7f0e95a2\") " pod="minio-dev/minio" Dec 01 19:53:28 crc kubenswrapper[4960]: I1201 19:53:28.801574 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.282215 4960 generic.go:334] "Generic (PLEG): container finished" podID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerID="c7bc951e1de7d985bc88fab794668e2baeac2ca48cfd8c351edb32e6bda7c34f" exitCode=0 Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.282425 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" event={"ID":"0f4799aa-209e-459d-a9e4-c2221c6a7e33","Type":"ContainerDied","Data":"c7bc951e1de7d985bc88fab794668e2baeac2ca48cfd8c351edb32e6bda7c34f"} Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.301896 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 01 19:53:29 crc kubenswrapper[4960]: W1201 19:53:29.311024 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcd5c18b_c010_4158_9935_033a7f0e95a2.slice/crio-a6823b95dad1c4f5883c9a08cf034dea88ef6dae86d59f873686c7d39378912f WatchSource:0}: Error finding container a6823b95dad1c4f5883c9a08cf034dea88ef6dae86d59f873686c7d39378912f: Status 404 returned error can't find the container with id a6823b95dad1c4f5883c9a08cf034dea88ef6dae86d59f873686c7d39378912f Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.589470 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.590767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.600754 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.699696 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.699924 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpxpn\" (UniqueName: \"kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.700087 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.801626 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.801713 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.801773 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpxpn\" (UniqueName: \"kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.802329 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.802384 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.824681 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpxpn\" (UniqueName: \"kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn\") pod \"redhat-operators-t5xzn\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:29 crc kubenswrapper[4960]: I1201 19:53:29.905956 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.144537 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.292049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fcd5c18b-c010-4158-9935-033a7f0e95a2","Type":"ContainerStarted","Data":"a6823b95dad1c4f5883c9a08cf034dea88ef6dae86d59f873686c7d39378912f"} Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.294222 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerStarted","Data":"f315b908a35714896dfd99fbf62641b834b2290375a566651051b9b6a7787c1c"} Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.294296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerStarted","Data":"0a28a3bb5b58711a5adfbeb7ef5eb90bfc382b4641f7366410b0d63e2edda829"} Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.304447 4960 generic.go:334] "Generic (PLEG): container finished" podID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerID="dc0909dadcf8a0a4740499f0968c8686d2759bee18731fbbfd14ec561c369b7d" exitCode=0 Dec 01 19:53:30 crc kubenswrapper[4960]: I1201 19:53:30.304544 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" event={"ID":"0f4799aa-209e-459d-a9e4-c2221c6a7e33","Type":"ContainerDied","Data":"dc0909dadcf8a0a4740499f0968c8686d2759bee18731fbbfd14ec561c369b7d"} Dec 01 19:53:31 crc kubenswrapper[4960]: I1201 19:53:31.315568 4960 generic.go:334] "Generic (PLEG): container finished" podID="87189451-856a-468f-bce5-7a0d5657f89d" containerID="f315b908a35714896dfd99fbf62641b834b2290375a566651051b9b6a7787c1c" exitCode=0 Dec 01 19:53:31 crc kubenswrapper[4960]: I1201 19:53:31.315767 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerDied","Data":"f315b908a35714896dfd99fbf62641b834b2290375a566651051b9b6a7787c1c"} Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.325060 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" event={"ID":"0f4799aa-209e-459d-a9e4-c2221c6a7e33","Type":"ContainerDied","Data":"a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4"} Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.325105 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3508b43cac184c6e7963912a9741aaee61018afae55caa9195f24faefc95bf4" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.331562 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.340460 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util\") pod \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.340572 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7blr\" (UniqueName: \"kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr\") pod \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.340632 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle\") pod \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\" (UID: \"0f4799aa-209e-459d-a9e4-c2221c6a7e33\") " Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.341695 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle" (OuterVolumeSpecName: "bundle") pod "0f4799aa-209e-459d-a9e4-c2221c6a7e33" (UID: "0f4799aa-209e-459d-a9e4-c2221c6a7e33"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.342676 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.349162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr" (OuterVolumeSpecName: "kube-api-access-j7blr") pod "0f4799aa-209e-459d-a9e4-c2221c6a7e33" (UID: "0f4799aa-209e-459d-a9e4-c2221c6a7e33"). InnerVolumeSpecName "kube-api-access-j7blr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.443475 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7blr\" (UniqueName: \"kubernetes.io/projected/0f4799aa-209e-459d-a9e4-c2221c6a7e33-kube-api-access-j7blr\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.589025 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util" (OuterVolumeSpecName: "util") pod "0f4799aa-209e-459d-a9e4-c2221c6a7e33" (UID: "0f4799aa-209e-459d-a9e4-c2221c6a7e33"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:32 crc kubenswrapper[4960]: I1201 19:53:32.646311 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0f4799aa-209e-459d-a9e4-c2221c6a7e33-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:33 crc kubenswrapper[4960]: I1201 19:53:33.329959 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf" Dec 01 19:53:34 crc kubenswrapper[4960]: I1201 19:53:34.341248 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"fcd5c18b-c010-4158-9935-033a7f0e95a2","Type":"ContainerStarted","Data":"3313f8dcac0702ee79f39ca5ef3c6e2768adece243bd29604ce776469213f4d7"} Dec 01 19:53:34 crc kubenswrapper[4960]: I1201 19:53:34.345269 4960 generic.go:334] "Generic (PLEG): container finished" podID="87189451-856a-468f-bce5-7a0d5657f89d" containerID="195da17638dbf3b83431220d36bdddee62a515e593503e20610b090b3bf2646e" exitCode=0 Dec 01 19:53:34 crc kubenswrapper[4960]: I1201 19:53:34.345371 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerDied","Data":"195da17638dbf3b83431220d36bdddee62a515e593503e20610b090b3bf2646e"} Dec 01 19:53:34 crc kubenswrapper[4960]: I1201 19:53:34.363725 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=5.506214851 podStartE2EDuration="9.363684885s" podCreationTimestamp="2025-12-01 19:53:25 +0000 UTC" firstStartedPulling="2025-12-01 19:53:29.322559739 +0000 UTC m=+844.610051418" lastFinishedPulling="2025-12-01 19:53:33.180029773 +0000 UTC m=+848.467521452" observedRunningTime="2025-12-01 19:53:34.363584782 +0000 UTC m=+849.651076491" watchObservedRunningTime="2025-12-01 19:53:34.363684885 +0000 UTC m=+849.651176594" Dec 01 19:53:35 crc kubenswrapper[4960]: I1201 19:53:35.357830 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerStarted","Data":"c754e4aa008da38f8a489a30a0da6533b5e525af7056fb88a6f2005ca7e49c68"} Dec 01 19:53:35 crc kubenswrapper[4960]: I1201 19:53:35.386687 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t5xzn" podStartSLOduration=2.90080236 podStartE2EDuration="6.386661618s" podCreationTimestamp="2025-12-01 19:53:29 +0000 UTC" firstStartedPulling="2025-12-01 19:53:31.371169356 +0000 UTC m=+846.658661025" lastFinishedPulling="2025-12-01 19:53:34.857028604 +0000 UTC m=+850.144520283" observedRunningTime="2025-12-01 19:53:35.384353836 +0000 UTC m=+850.671845535" watchObservedRunningTime="2025-12-01 19:53:35.386661618 +0000 UTC m=+850.674153287" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.654449 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62"] Dec 01 19:53:39 crc kubenswrapper[4960]: E1201 19:53:39.655393 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="extract" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.655417 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="extract" Dec 01 19:53:39 crc kubenswrapper[4960]: E1201 19:53:39.655457 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="util" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.655469 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="util" Dec 01 19:53:39 crc kubenswrapper[4960]: E1201 19:53:39.655486 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="pull" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.655497 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="pull" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.655680 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" containerName="extract" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.657086 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.659515 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.671561 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62"] Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.751275 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.751355 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqgg2\" (UniqueName: \"kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.751404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.853187 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.853261 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.853310 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqgg2\" (UniqueName: \"kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.853892 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.854219 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.875423 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqgg2\" (UniqueName: \"kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2\") pod \"03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.906790 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:39 crc kubenswrapper[4960]: I1201 19:53:39.907194 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:40 crc kubenswrapper[4960]: I1201 19:53:40.011348 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:40 crc kubenswrapper[4960]: I1201 19:53:40.279397 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62"] Dec 01 19:53:40 crc kubenswrapper[4960]: I1201 19:53:40.393020 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" event={"ID":"1b510133-6226-41f2-916c-61b8f0130f86","Type":"ContainerStarted","Data":"3bb8c8f22c9b504be935719fc07a62d5226b066ea15390ebb6d584af53a2a040"} Dec 01 19:53:40 crc kubenswrapper[4960]: I1201 19:53:40.952963 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t5xzn" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="registry-server" probeResult="failure" output=< Dec 01 19:53:40 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 19:53:40 crc kubenswrapper[4960]: > Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.105151 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb"] Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.106493 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.110172 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.116966 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.117180 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.117338 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.117509 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-ng27b" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.117623 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.127665 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb"] Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.275053 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.275297 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5qf5\" (UniqueName: \"kubernetes.io/projected/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-kube-api-access-t5qf5\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.275351 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-webhook-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.275439 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-apiservice-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.275503 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-manager-config\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.377079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5qf5\" (UniqueName: \"kubernetes.io/projected/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-kube-api-access-t5qf5\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.377167 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-webhook-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.377210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-apiservice-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.377281 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-manager-config\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.378850 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-manager-config\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.378995 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.384948 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-apiservice-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.385026 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-webhook-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.387804 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.407767 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5qf5\" (UniqueName: \"kubernetes.io/projected/40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b-kube-api-access-t5qf5\") pod \"loki-operator-controller-manager-6dfbdd7cb8-tcjcb\" (UID: \"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b\") " pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.428212 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b510133-6226-41f2-916c-61b8f0130f86" containerID="caa3ee1e590d89841e8b868a9a2c2aeeb638bf3d383bfd4cb3248f2450eece7f" exitCode=0 Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.428268 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" event={"ID":"1b510133-6226-41f2-916c-61b8f0130f86","Type":"ContainerDied","Data":"caa3ee1e590d89841e8b868a9a2c2aeeb638bf3d383bfd4cb3248f2450eece7f"} Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.462604 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:41 crc kubenswrapper[4960]: I1201 19:53:41.656808 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb"] Dec 01 19:53:42 crc kubenswrapper[4960]: I1201 19:53:42.439962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" event={"ID":"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b","Type":"ContainerStarted","Data":"754e81fbf67d21d4de32b70762bea52fe6015b82acf333050cf8e46333db713c"} Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.451870 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b510133-6226-41f2-916c-61b8f0130f86" containerID="1016e2782ebac5b8c32f8a5042c2e901e149b677005c278b4a8612728e0d0022" exitCode=0 Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.452149 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" event={"ID":"1b510133-6226-41f2-916c-61b8f0130f86","Type":"ContainerDied","Data":"1016e2782ebac5b8c32f8a5042c2e901e149b677005c278b4a8612728e0d0022"} Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.870844 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.870922 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.870975 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.871752 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 19:53:43 crc kubenswrapper[4960]: I1201 19:53:43.871828 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d" gracePeriod=600 Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.464814 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d" exitCode=0 Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.465175 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d"} Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.465360 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09"} Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.465392 4960 scope.go:117] "RemoveContainer" containerID="117dcdc4e75073ecdf5ccc0361aa47bf5b8771625110f4ebc1d8f3e3848e8b6b" Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.479389 4960 generic.go:334] "Generic (PLEG): container finished" podID="1b510133-6226-41f2-916c-61b8f0130f86" containerID="9977357e08d68d2be3da0a3f6b1aff32c248eed26d0eda8802531e5d43acc1bc" exitCode=0 Dec 01 19:53:44 crc kubenswrapper[4960]: I1201 19:53:44.479454 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" event={"ID":"1b510133-6226-41f2-916c-61b8f0130f86","Type":"ContainerDied","Data":"9977357e08d68d2be3da0a3f6b1aff32c248eed26d0eda8802531e5d43acc1bc"} Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.027719 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.074846 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util\") pod \"1b510133-6226-41f2-916c-61b8f0130f86\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.075439 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle\") pod \"1b510133-6226-41f2-916c-61b8f0130f86\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.075546 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqgg2\" (UniqueName: \"kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2\") pod \"1b510133-6226-41f2-916c-61b8f0130f86\" (UID: \"1b510133-6226-41f2-916c-61b8f0130f86\") " Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.077582 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle" (OuterVolumeSpecName: "bundle") pod "1b510133-6226-41f2-916c-61b8f0130f86" (UID: "1b510133-6226-41f2-916c-61b8f0130f86"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.085202 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2" (OuterVolumeSpecName: "kube-api-access-xqgg2") pod "1b510133-6226-41f2-916c-61b8f0130f86" (UID: "1b510133-6226-41f2-916c-61b8f0130f86"). InnerVolumeSpecName "kube-api-access-xqgg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.102100 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util" (OuterVolumeSpecName: "util") pod "1b510133-6226-41f2-916c-61b8f0130f86" (UID: "1b510133-6226-41f2-916c-61b8f0130f86"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.177048 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqgg2\" (UniqueName: \"kubernetes.io/projected/1b510133-6226-41f2-916c-61b8f0130f86-kube-api-access-xqgg2\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.177101 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.177133 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1b510133-6226-41f2-916c-61b8f0130f86-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.500318 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.500334 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62" event={"ID":"1b510133-6226-41f2-916c-61b8f0130f86","Type":"ContainerDied","Data":"3bb8c8f22c9b504be935719fc07a62d5226b066ea15390ebb6d584af53a2a040"} Dec 01 19:53:46 crc kubenswrapper[4960]: I1201 19:53:46.500402 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bb8c8f22c9b504be935719fc07a62d5226b066ea15390ebb6d584af53a2a040" Dec 01 19:53:47 crc kubenswrapper[4960]: I1201 19:53:47.507378 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" event={"ID":"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b","Type":"ContainerStarted","Data":"2b773d5ee3f019f01de7fd6da7be8ecaf1828751965a91b1a79189eabb751347"} Dec 01 19:53:49 crc kubenswrapper[4960]: I1201 19:53:49.944708 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:49 crc kubenswrapper[4960]: I1201 19:53:49.984826 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.384251 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.385031 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t5xzn" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="registry-server" containerID="cri-o://c754e4aa008da38f8a489a30a0da6533b5e525af7056fb88a6f2005ca7e49c68" gracePeriod=2 Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.561578 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" event={"ID":"40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b","Type":"ContainerStarted","Data":"d9d1bca1ea12fbe285d2b831827652788a5b5d1262ba565f2bfea5a41f1f0734"} Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.562255 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.565232 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.570640 4960 generic.go:334] "Generic (PLEG): container finished" podID="87189451-856a-468f-bce5-7a0d5657f89d" containerID="c754e4aa008da38f8a489a30a0da6533b5e525af7056fb88a6f2005ca7e49c68" exitCode=0 Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.570779 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerDied","Data":"c754e4aa008da38f8a489a30a0da6533b5e525af7056fb88a6f2005ca7e49c68"} Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.604006 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-6dfbdd7cb8-tcjcb" podStartSLOduration=1.270675525 podStartE2EDuration="12.603976679s" podCreationTimestamp="2025-12-01 19:53:41 +0000 UTC" firstStartedPulling="2025-12-01 19:53:41.670619689 +0000 UTC m=+856.958111358" lastFinishedPulling="2025-12-01 19:53:53.003920833 +0000 UTC m=+868.291412512" observedRunningTime="2025-12-01 19:53:53.60370595 +0000 UTC m=+868.891197659" watchObservedRunningTime="2025-12-01 19:53:53.603976679 +0000 UTC m=+868.891468388" Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.837129 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.995381 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpxpn\" (UniqueName: \"kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn\") pod \"87189451-856a-468f-bce5-7a0d5657f89d\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.995545 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content\") pod \"87189451-856a-468f-bce5-7a0d5657f89d\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.995628 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities\") pod \"87189451-856a-468f-bce5-7a0d5657f89d\" (UID: \"87189451-856a-468f-bce5-7a0d5657f89d\") " Dec 01 19:53:53 crc kubenswrapper[4960]: I1201 19:53:53.996957 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities" (OuterVolumeSpecName: "utilities") pod "87189451-856a-468f-bce5-7a0d5657f89d" (UID: "87189451-856a-468f-bce5-7a0d5657f89d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.016458 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn" (OuterVolumeSpecName: "kube-api-access-xpxpn") pod "87189451-856a-468f-bce5-7a0d5657f89d" (UID: "87189451-856a-468f-bce5-7a0d5657f89d"). InnerVolumeSpecName "kube-api-access-xpxpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.096792 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpxpn\" (UniqueName: \"kubernetes.io/projected/87189451-856a-468f-bce5-7a0d5657f89d-kube-api-access-xpxpn\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.097075 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.103430 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87189451-856a-468f-bce5-7a0d5657f89d" (UID: "87189451-856a-468f-bce5-7a0d5657f89d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.198504 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87189451-856a-468f-bce5-7a0d5657f89d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.584898 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5xzn" event={"ID":"87189451-856a-468f-bce5-7a0d5657f89d","Type":"ContainerDied","Data":"0a28a3bb5b58711a5adfbeb7ef5eb90bfc382b4641f7366410b0d63e2edda829"} Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.585299 4960 scope.go:117] "RemoveContainer" containerID="c754e4aa008da38f8a489a30a0da6533b5e525af7056fb88a6f2005ca7e49c68" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.585176 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5xzn" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.619348 4960 scope.go:117] "RemoveContainer" containerID="195da17638dbf3b83431220d36bdddee62a515e593503e20610b090b3bf2646e" Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.623157 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.634897 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t5xzn"] Dec 01 19:53:54 crc kubenswrapper[4960]: I1201 19:53:54.652597 4960 scope.go:117] "RemoveContainer" containerID="f315b908a35714896dfd99fbf62641b834b2290375a566651051b9b6a7787c1c" Dec 01 19:53:55 crc kubenswrapper[4960]: I1201 19:53:55.332743 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87189451-856a-468f-bce5-7a0d5657f89d" path="/var/lib/kubelet/pods/87189451-856a-468f-bce5-7a0d5657f89d/volumes" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.199176 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200237 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="registry-server" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200251 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="registry-server" Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200270 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="extract-utilities" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200276 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="extract-utilities" Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200289 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="extract" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200295 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="extract" Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200307 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="extract-content" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200313 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="extract-content" Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200320 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="util" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200327 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="util" Dec 01 19:54:23 crc kubenswrapper[4960]: E1201 19:54:23.200335 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="pull" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200340 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="pull" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200455 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="87189451-856a-468f-bce5-7a0d5657f89d" containerName="registry-server" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.200470 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b510133-6226-41f2-916c-61b8f0130f86" containerName="extract" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.201550 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.222734 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5j65\" (UniqueName: \"kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.227664 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.227954 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.261237 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.329492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5j65\" (UniqueName: \"kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.329546 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.329577 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.330178 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.330222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.355954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5j65\" (UniqueName: \"kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65\") pod \"certified-operators-kzzpq\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.554088 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:23 crc kubenswrapper[4960]: I1201 19:54:23.815798 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:24 crc kubenswrapper[4960]: I1201 19:54:24.811590 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerID="e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31" exitCode=0 Dec 01 19:54:24 crc kubenswrapper[4960]: I1201 19:54:24.811830 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerDied","Data":"e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31"} Dec 01 19:54:24 crc kubenswrapper[4960]: I1201 19:54:24.814464 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerStarted","Data":"e3d933efdc4b06ff6306b21ea889f6ca6ab490fa92479d8777b1d68475a8ae91"} Dec 01 19:54:26 crc kubenswrapper[4960]: I1201 19:54:26.839609 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerID="e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45" exitCode=0 Dec 01 19:54:26 crc kubenswrapper[4960]: I1201 19:54:26.840620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerDied","Data":"e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45"} Dec 01 19:54:27 crc kubenswrapper[4960]: I1201 19:54:27.851397 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerStarted","Data":"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68"} Dec 01 19:54:27 crc kubenswrapper[4960]: I1201 19:54:27.870087 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kzzpq" podStartSLOduration=2.125402197 podStartE2EDuration="4.870049275s" podCreationTimestamp="2025-12-01 19:54:23 +0000 UTC" firstStartedPulling="2025-12-01 19:54:24.815967218 +0000 UTC m=+900.103458897" lastFinishedPulling="2025-12-01 19:54:27.560614286 +0000 UTC m=+902.848105975" observedRunningTime="2025-12-01 19:54:27.868987772 +0000 UTC m=+903.156479441" watchObservedRunningTime="2025-12-01 19:54:27.870049275 +0000 UTC m=+903.157540944" Dec 01 19:54:33 crc kubenswrapper[4960]: I1201 19:54:33.554412 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:33 crc kubenswrapper[4960]: I1201 19:54:33.555340 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:33 crc kubenswrapper[4960]: I1201 19:54:33.644672 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:33 crc kubenswrapper[4960]: I1201 19:54:33.948918 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:33 crc kubenswrapper[4960]: I1201 19:54:33.995474 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:35 crc kubenswrapper[4960]: I1201 19:54:35.926713 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kzzpq" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="registry-server" containerID="cri-o://fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68" gracePeriod=2 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.131842 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.138664 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/03c6e0f8bd928fdcaaf530d547155f7eef49635d3e29724a094c0ab694zgt62"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.146236 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.151906 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/142e5edc705b0443a978f15b9d74db4e11d2db1d26a61e7f8c9e49e303ws2mf"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.165499 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.172030 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210swh8k"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.176239 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.176524 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vhhcs" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="registry-server" containerID="cri-o://afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80" gracePeriod=30 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.183132 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.183483 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xct8q" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="registry-server" containerID="cri-o://ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7" gracePeriod=30 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.203293 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.203579 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" podUID="544fa4c7-f389-4765-8880-d6fa67065167" containerName="marketplace-operator" containerID="cri-o://69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000" gracePeriod=30 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.215827 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.216059 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-69sk4" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="registry-server" containerID="cri-o://4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324" gracePeriod=30 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.226051 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.226371 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qwd6x" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="registry-server" containerID="cri-o://bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e" gracePeriod=30 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.231219 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsjbd"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.237528 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.247853 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsjbd"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.331698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.332524 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psq8v\" (UniqueName: \"kubernetes.io/projected/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-kube-api-access-psq8v\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.332551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.433700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.433877 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psq8v\" (UniqueName: \"kubernetes.io/projected/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-kube-api-access-psq8v\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.433899 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.437133 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.442584 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.456074 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psq8v\" (UniqueName: \"kubernetes.io/projected/a968fcf3-9935-4cfe-abdc-82c62d9b3b12-kube-api-access-psq8v\") pod \"marketplace-operator-79b997595-qsjbd\" (UID: \"a968fcf3-9935-4cfe-abdc-82c62d9b3b12\") " pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.615189 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.616537 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.621043 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.654877 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.671444 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.680672 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.705247 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747555 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content\") pod \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747665 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities\") pod \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747722 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjbnf\" (UniqueName: \"kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf\") pod \"0237c051-807d-4c6c-8f72-71cd35f2a8df\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747788 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content\") pod \"8b458ace-89d3-4a6f-950e-a7f54f90f340\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747821 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities\") pod \"8b458ace-89d3-4a6f-950e-a7f54f90f340\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747853 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities\") pod \"0237c051-807d-4c6c-8f72-71cd35f2a8df\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747892 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5j65\" (UniqueName: \"kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65\") pod \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\" (UID: \"b6b35552-6680-4a88-8269-9c3e6da1ffb8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747931 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content\") pod \"0237c051-807d-4c6c-8f72-71cd35f2a8df\" (UID: \"0237c051-807d-4c6c-8f72-71cd35f2a8df\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.747976 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzdb2\" (UniqueName: \"kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2\") pod \"8b458ace-89d3-4a6f-950e-a7f54f90f340\" (UID: \"8b458ace-89d3-4a6f-950e-a7f54f90f340\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.748206 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r22d\" (UniqueName: \"kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.748323 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.748353 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.753677 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities" (OuterVolumeSpecName: "utilities") pod "0237c051-807d-4c6c-8f72-71cd35f2a8df" (UID: "0237c051-807d-4c6c-8f72-71cd35f2a8df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.754520 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities" (OuterVolumeSpecName: "utilities") pod "b6b35552-6680-4a88-8269-9c3e6da1ffb8" (UID: "b6b35552-6680-4a88-8269-9c3e6da1ffb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.756276 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2" (OuterVolumeSpecName: "kube-api-access-mzdb2") pod "8b458ace-89d3-4a6f-950e-a7f54f90f340" (UID: "8b458ace-89d3-4a6f-950e-a7f54f90f340"). InnerVolumeSpecName "kube-api-access-mzdb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.759302 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf" (OuterVolumeSpecName: "kube-api-access-gjbnf") pod "0237c051-807d-4c6c-8f72-71cd35f2a8df" (UID: "0237c051-807d-4c6c-8f72-71cd35f2a8df"). InnerVolumeSpecName "kube-api-access-gjbnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.759528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65" (OuterVolumeSpecName: "kube-api-access-h5j65") pod "b6b35552-6680-4a88-8269-9c3e6da1ffb8" (UID: "b6b35552-6680-4a88-8269-9c3e6da1ffb8"). InnerVolumeSpecName "kube-api-access-h5j65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.761541 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.761528 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities" (OuterVolumeSpecName: "utilities") pod "8b458ace-89d3-4a6f-950e-a7f54f90f340" (UID: "8b458ace-89d3-4a6f-950e-a7f54f90f340"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.795620 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.810651 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.844209 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0237c051-807d-4c6c-8f72-71cd35f2a8df" (UID: "0237c051-807d-4c6c-8f72-71cd35f2a8df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.846194 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6b35552-6680-4a88-8269-9c3e6da1ffb8" (UID: "b6b35552-6680-4a88-8269-9c3e6da1ffb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855558 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content\") pod \"05309d39-9cfb-430f-b69f-9ace1978796f\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855628 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content\") pod \"15c5f8a1-4662-457f-962a-c7ba801056e8\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855665 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics\") pod \"544fa4c7-f389-4765-8880-d6fa67065167\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855705 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85q84\" (UniqueName: \"kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84\") pod \"544fa4c7-f389-4765-8880-d6fa67065167\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855739 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities\") pod \"15c5f8a1-4662-457f-962a-c7ba801056e8\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855762 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca\") pod \"544fa4c7-f389-4765-8880-d6fa67065167\" (UID: \"544fa4c7-f389-4765-8880-d6fa67065167\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855781 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8p55\" (UniqueName: \"kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55\") pod \"05309d39-9cfb-430f-b69f-9ace1978796f\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855801 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities\") pod \"05309d39-9cfb-430f-b69f-9ace1978796f\" (UID: \"05309d39-9cfb-430f-b69f-9ace1978796f\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855835 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp8zf\" (UniqueName: \"kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf\") pod \"15c5f8a1-4662-457f-962a-c7ba801056e8\" (UID: \"15c5f8a1-4662-457f-962a-c7ba801056e8\") " Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.855984 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856022 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r22d\" (UniqueName: \"kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856078 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856131 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856142 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjbnf\" (UniqueName: \"kubernetes.io/projected/0237c051-807d-4c6c-8f72-71cd35f2a8df-kube-api-access-gjbnf\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856154 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856162 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856172 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5j65\" (UniqueName: \"kubernetes.io/projected/b6b35552-6680-4a88-8269-9c3e6da1ffb8-kube-api-access-h5j65\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856181 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0237c051-807d-4c6c-8f72-71cd35f2a8df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856191 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzdb2\" (UniqueName: \"kubernetes.io/projected/8b458ace-89d3-4a6f-950e-a7f54f90f340-kube-api-access-mzdb2\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856199 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b35552-6680-4a88-8269-9c3e6da1ffb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.856763 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.859168 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b458ace-89d3-4a6f-950e-a7f54f90f340" (UID: "8b458ace-89d3-4a6f-950e-a7f54f90f340"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.860454 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities" (OuterVolumeSpecName: "utilities") pod "05309d39-9cfb-430f-b69f-9ace1978796f" (UID: "05309d39-9cfb-430f-b69f-9ace1978796f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.860943 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.861952 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities" (OuterVolumeSpecName: "utilities") pod "15c5f8a1-4662-457f-962a-c7ba801056e8" (UID: "15c5f8a1-4662-457f-962a-c7ba801056e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.862064 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "544fa4c7-f389-4765-8880-d6fa67065167" (UID: "544fa4c7-f389-4765-8880-d6fa67065167"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.865525 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55" (OuterVolumeSpecName: "kube-api-access-h8p55") pod "05309d39-9cfb-430f-b69f-9ace1978796f" (UID: "05309d39-9cfb-430f-b69f-9ace1978796f"). InnerVolumeSpecName "kube-api-access-h8p55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.865903 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf" (OuterVolumeSpecName: "kube-api-access-zp8zf") pod "15c5f8a1-4662-457f-962a-c7ba801056e8" (UID: "15c5f8a1-4662-457f-962a-c7ba801056e8"). InnerVolumeSpecName "kube-api-access-zp8zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.867049 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84" (OuterVolumeSpecName: "kube-api-access-85q84") pod "544fa4c7-f389-4765-8880-d6fa67065167" (UID: "544fa4c7-f389-4765-8880-d6fa67065167"). InnerVolumeSpecName "kube-api-access-85q84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.869523 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "544fa4c7-f389-4765-8880-d6fa67065167" (UID: "544fa4c7-f389-4765-8880-d6fa67065167"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.882472 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05309d39-9cfb-430f-b69f-9ace1978796f" (UID: "05309d39-9cfb-430f-b69f-9ace1978796f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.882827 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r22d\" (UniqueName: \"kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d\") pod \"community-operators-9wcx2\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.945577 4960 generic.go:334] "Generic (PLEG): container finished" podID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerID="afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.945658 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerDied","Data":"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.945693 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vhhcs" event={"ID":"0237c051-807d-4c6c-8f72-71cd35f2a8df","Type":"ContainerDied","Data":"7fbd601fcba81ec6483beb340f211e946e742bd3ca05ac43406a8329b52fba99"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.945713 4960 scope.go:117] "RemoveContainer" containerID="afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.945767 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vhhcs" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.946183 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qsjbd"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.949250 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerID="fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.949304 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerDied","Data":"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.949323 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kzzpq" event={"ID":"b6b35552-6680-4a88-8269-9c3e6da1ffb8","Type":"ContainerDied","Data":"e3d933efdc4b06ff6306b21ea889f6ca6ab490fa92479d8777b1d68475a8ae91"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.949390 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kzzpq" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.956479 4960 generic.go:334] "Generic (PLEG): container finished" podID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerID="bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.956532 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerDied","Data":"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.956556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qwd6x" event={"ID":"15c5f8a1-4662-457f-962a-c7ba801056e8","Type":"ContainerDied","Data":"45a1f2fab9838c94bf25104e0d2032ab99c941fcae0cb6e3758e7330784640d4"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.956621 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qwd6x" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957093 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957132 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957143 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8p55\" (UniqueName: \"kubernetes.io/projected/05309d39-9cfb-430f-b69f-9ace1978796f-kube-api-access-h8p55\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957153 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957163 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp8zf\" (UniqueName: \"kubernetes.io/projected/15c5f8a1-4662-457f-962a-c7ba801056e8-kube-api-access-zp8zf\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957171 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b458ace-89d3-4a6f-950e-a7f54f90f340-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957179 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05309d39-9cfb-430f-b69f-9ace1978796f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957188 4960 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/544fa4c7-f389-4765-8880-d6fa67065167-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.957195 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85q84\" (UniqueName: \"kubernetes.io/projected/544fa4c7-f389-4765-8880-d6fa67065167-kube-api-access-85q84\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.961928 4960 generic.go:334] "Generic (PLEG): container finished" podID="05309d39-9cfb-430f-b69f-9ace1978796f" containerID="4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.961976 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerDied","Data":"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.962036 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69sk4" event={"ID":"05309d39-9cfb-430f-b69f-9ace1978796f","Type":"ContainerDied","Data":"64af68bbb5cfd59872cd424d3da698e86993d6916ea97b2cde83c6c1c043193f"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.962058 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69sk4" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.963607 4960 generic.go:334] "Generic (PLEG): container finished" podID="544fa4c7-f389-4765-8880-d6fa67065167" containerID="69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.963676 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.963661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" event={"ID":"544fa4c7-f389-4765-8880-d6fa67065167","Type":"ContainerDied","Data":"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.963791 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wckns" event={"ID":"544fa4c7-f389-4765-8880-d6fa67065167","Type":"ContainerDied","Data":"f56e66b2f130446730f77050f680bc879f0aca7419f7ebcf4140fc1b58e4b85f"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.969444 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerID="ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7" exitCode=0 Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.969488 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerDied","Data":"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.969508 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xct8q" event={"ID":"8b458ace-89d3-4a6f-950e-a7f54f90f340","Type":"ContainerDied","Data":"cca32d5375867dbdf3fd0ec2862c1658ce3f23783e1c1bcde9e863b1587cec8d"} Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.969547 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xct8q" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.981129 4960 scope.go:117] "RemoveContainer" containerID="75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.985238 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.988446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.988987 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kzzpq"] Dec 01 19:54:36 crc kubenswrapper[4960]: I1201 19:54:36.996066 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15c5f8a1-4662-457f-962a-c7ba801056e8" (UID: "15c5f8a1-4662-457f-962a-c7ba801056e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.001414 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.004445 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vhhcs"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.024843 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.028296 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-69sk4"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.030918 4960 scope.go:117] "RemoveContainer" containerID="6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.041105 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.047820 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wckns"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.056995 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.058103 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15c5f8a1-4662-457f-962a-c7ba801056e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.063331 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xct8q"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.067873 4960 scope.go:117] "RemoveContainer" containerID="afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.068491 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80\": container with ID starting with afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80 not found: ID does not exist" containerID="afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.068580 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80"} err="failed to get container status \"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80\": rpc error: code = NotFound desc = could not find container \"afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80\": container with ID starting with afc7773a86c9613cf7930168f6355e34846464fdd74d1be1a5c8b538b3c8da80 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.068685 4960 scope.go:117] "RemoveContainer" containerID="75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.069019 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778\": container with ID starting with 75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778 not found: ID does not exist" containerID="75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.069097 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778"} err="failed to get container status \"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778\": rpc error: code = NotFound desc = could not find container \"75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778\": container with ID starting with 75baab9e7cdf127f14f2c07bd029771ccd2aa6adba977d8d7ac28c4bc11d6778 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.069177 4960 scope.go:117] "RemoveContainer" containerID="6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.069534 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1\": container with ID starting with 6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1 not found: ID does not exist" containerID="6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.069654 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1"} err="failed to get container status \"6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1\": rpc error: code = NotFound desc = could not find container \"6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1\": container with ID starting with 6b4d6724395cbe9180b097a7bbf5e019d77e531d77a5f2d9de2bd37d1de1e6b1 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.069719 4960 scope.go:117] "RemoveContainer" containerID="fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.135704 4960 scope.go:117] "RemoveContainer" containerID="e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.156757 4960 scope.go:117] "RemoveContainer" containerID="e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.186609 4960 scope.go:117] "RemoveContainer" containerID="fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.187271 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68\": container with ID starting with fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68 not found: ID does not exist" containerID="fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.187331 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68"} err="failed to get container status \"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68\": rpc error: code = NotFound desc = could not find container \"fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68\": container with ID starting with fd036f400280d85a5c754a178f6d016e7233712ce4e66cd79f47f9fda222bb68 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.187368 4960 scope.go:117] "RemoveContainer" containerID="e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.187808 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45\": container with ID starting with e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45 not found: ID does not exist" containerID="e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.187847 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45"} err="failed to get container status \"e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45\": rpc error: code = NotFound desc = could not find container \"e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45\": container with ID starting with e7ed8e477188be2826eef5e5eb65e857198cb18bcb1ef7f57b3cfba41ea0de45 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.187876 4960 scope.go:117] "RemoveContainer" containerID="e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.188630 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31\": container with ID starting with e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31 not found: ID does not exist" containerID="e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.188706 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31"} err="failed to get container status \"e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31\": rpc error: code = NotFound desc = could not find container \"e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31\": container with ID starting with e16c1aaf3221e4912224855d721b79f7872acda35687e7557b09b8891d64ea31 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.188762 4960 scope.go:117] "RemoveContainer" containerID="bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.213873 4960 scope.go:117] "RemoveContainer" containerID="683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.238157 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.243290 4960 scope.go:117] "RemoveContainer" containerID="eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df" Dec 01 19:54:37 crc kubenswrapper[4960]: W1201 19:54:37.249130 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb71eafc4_b189_4d2f_ad28_be0b1024d1c9.slice/crio-a2f167d3ae66a84e84da93254eec4603bccd210b2c69ed677110c83b5a2b9d18 WatchSource:0}: Error finding container a2f167d3ae66a84e84da93254eec4603bccd210b2c69ed677110c83b5a2b9d18: Status 404 returned error can't find the container with id a2f167d3ae66a84e84da93254eec4603bccd210b2c69ed677110c83b5a2b9d18 Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.263827 4960 scope.go:117] "RemoveContainer" containerID="bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.264427 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e\": container with ID starting with bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e not found: ID does not exist" containerID="bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.264462 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e"} err="failed to get container status \"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e\": rpc error: code = NotFound desc = could not find container \"bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e\": container with ID starting with bac227d5af3cc2fc3172f35c02f5cfe91989603ed277349ca88d086c69620c6e not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.264484 4960 scope.go:117] "RemoveContainer" containerID="683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.264917 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60\": container with ID starting with 683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60 not found: ID does not exist" containerID="683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.264936 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60"} err="failed to get container status \"683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60\": rpc error: code = NotFound desc = could not find container \"683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60\": container with ID starting with 683797ec1349b0f510ea000c7d85b5cd13dcb8e551d779f672fd2951b118aa60 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.264957 4960 scope.go:117] "RemoveContainer" containerID="eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.265743 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df\": container with ID starting with eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df not found: ID does not exist" containerID="eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.265767 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df"} err="failed to get container status \"eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df\": rpc error: code = NotFound desc = could not find container \"eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df\": container with ID starting with eaa1aaf39b4f7da32f96d5c0dee54212a9a30b60a015ea36dfde565bb29af9df not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.265783 4960 scope.go:117] "RemoveContainer" containerID="4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.290945 4960 scope.go:117] "RemoveContainer" containerID="a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.292037 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.298433 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qwd6x"] Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.307007 4960 scope.go:117] "RemoveContainer" containerID="9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.330867 4960 scope.go:117] "RemoveContainer" containerID="4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.331469 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324\": container with ID starting with 4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324 not found: ID does not exist" containerID="4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.331527 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324"} err="failed to get container status \"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324\": rpc error: code = NotFound desc = could not find container \"4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324\": container with ID starting with 4cedee468fb9090b1bcad82f270f99a3031db17cca5df5cad8bb9f498e03c324 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.331567 4960 scope.go:117] "RemoveContainer" containerID="a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.333283 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549\": container with ID starting with a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549 not found: ID does not exist" containerID="a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.333328 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549"} err="failed to get container status \"a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549\": rpc error: code = NotFound desc = could not find container \"a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549\": container with ID starting with a87f63d60d360af9adec96c41cf87148afab61c281ab33e1e3c4b2005738f549 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.333357 4960 scope.go:117] "RemoveContainer" containerID="9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.333726 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172\": container with ID starting with 9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172 not found: ID does not exist" containerID="9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.333757 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172"} err="failed to get container status \"9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172\": rpc error: code = NotFound desc = could not find container \"9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172\": container with ID starting with 9ff1f81900b2455a9024a459eda93e6cd66fb9cd784f700ad22d09403f7ae172 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.333774 4960 scope.go:117] "RemoveContainer" containerID="69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.333741 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" path="/var/lib/kubelet/pods/0237c051-807d-4c6c-8f72-71cd35f2a8df/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.334723 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="028186f6-ad65-47f2-be65-50d0e15dfc12" path="/var/lib/kubelet/pods/028186f6-ad65-47f2-be65-50d0e15dfc12/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.335408 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" path="/var/lib/kubelet/pods/05309d39-9cfb-430f-b69f-9ace1978796f/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.336597 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f4799aa-209e-459d-a9e4-c2221c6a7e33" path="/var/lib/kubelet/pods/0f4799aa-209e-459d-a9e4-c2221c6a7e33/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.337248 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" path="/var/lib/kubelet/pods/15c5f8a1-4662-457f-962a-c7ba801056e8/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.338493 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b510133-6226-41f2-916c-61b8f0130f86" path="/var/lib/kubelet/pods/1b510133-6226-41f2-916c-61b8f0130f86/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.339179 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="544fa4c7-f389-4765-8880-d6fa67065167" path="/var/lib/kubelet/pods/544fa4c7-f389-4765-8880-d6fa67065167/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.339620 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" path="/var/lib/kubelet/pods/8b458ace-89d3-4a6f-950e-a7f54f90f340/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.340695 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" path="/var/lib/kubelet/pods/b6b35552-6680-4a88-8269-9c3e6da1ffb8/volumes" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.351436 4960 scope.go:117] "RemoveContainer" containerID="69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.351882 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000\": container with ID starting with 69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000 not found: ID does not exist" containerID="69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.351949 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000"} err="failed to get container status \"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000\": rpc error: code = NotFound desc = could not find container \"69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000\": container with ID starting with 69e3449ee47009467d6e003e44483c82ce81b5f3af65982d18066615ee312000 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.351993 4960 scope.go:117] "RemoveContainer" containerID="ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.368684 4960 scope.go:117] "RemoveContainer" containerID="0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.385179 4960 scope.go:117] "RemoveContainer" containerID="acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.408990 4960 scope.go:117] "RemoveContainer" containerID="ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.409398 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7\": container with ID starting with ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7 not found: ID does not exist" containerID="ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.409430 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7"} err="failed to get container status \"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7\": rpc error: code = NotFound desc = could not find container \"ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7\": container with ID starting with ccef55ed58fa626984efee89e790314833803b0face29866dba10093097058a7 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.409456 4960 scope.go:117] "RemoveContainer" containerID="0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.410323 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd\": container with ID starting with 0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd not found: ID does not exist" containerID="0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.410382 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd"} err="failed to get container status \"0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd\": rpc error: code = NotFound desc = could not find container \"0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd\": container with ID starting with 0e530676fc47833729174bdb101cba95da2943c08a0f85b61a9828306bd8eafd not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.410423 4960 scope.go:117] "RemoveContainer" containerID="acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54" Dec 01 19:54:37 crc kubenswrapper[4960]: E1201 19:54:37.410833 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54\": container with ID starting with acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54 not found: ID does not exist" containerID="acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.410878 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54"} err="failed to get container status \"acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54\": rpc error: code = NotFound desc = could not find container \"acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54\": container with ID starting with acfcfcf3b4971b6192331842fd167571e8f36b743f949c54f5c4afaba38ffa54 not found: ID does not exist" Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.977368 4960 generic.go:334] "Generic (PLEG): container finished" podID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerID="94f1cf8a1915f28ee6149f17ad6448760a5bc6a827ed7dab9c0924255239fa72" exitCode=0 Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.978244 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerDied","Data":"94f1cf8a1915f28ee6149f17ad6448760a5bc6a827ed7dab9c0924255239fa72"} Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.978349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerStarted","Data":"a2f167d3ae66a84e84da93254eec4603bccd210b2c69ed677110c83b5a2b9d18"} Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.985962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" event={"ID":"a968fcf3-9935-4cfe-abdc-82c62d9b3b12","Type":"ContainerStarted","Data":"1159bb3b601b8be0d9a5cfbf0a96f92cbddee292b22082c4d96508df707cad1f"} Dec 01 19:54:37 crc kubenswrapper[4960]: I1201 19:54:37.986025 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" event={"ID":"a968fcf3-9935-4cfe-abdc-82c62d9b3b12","Type":"ContainerStarted","Data":"578e557a755ae16aef286aab76fef2807f8a66640e269a5eb3639e63cc0cc7fd"} Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.023954 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" podStartSLOduration=2.023921161 podStartE2EDuration="2.023921161s" podCreationTimestamp="2025-12-01 19:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:54:38.019086061 +0000 UTC m=+913.306577770" watchObservedRunningTime="2025-12-01 19:54:38.023921161 +0000 UTC m=+913.311412870" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394175 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jcfpr"] Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394604 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394624 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394642 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394654 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394693 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394704 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394723 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394733 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394746 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394783 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394803 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.394812 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.394826 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395022 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395042 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395050 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395060 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395068 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395104 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="544fa4c7-f389-4765-8880-d6fa67065167" containerName="marketplace-operator" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395150 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="544fa4c7-f389-4765-8880-d6fa67065167" containerName="marketplace-operator" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395198 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395232 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395247 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395255 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395269 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395277 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="extract-content" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395295 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395303 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395314 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395323 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: E1201 19:54:38.395334 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395342 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="extract-utilities" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395687 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="544fa4c7-f389-4765-8880-d6fa67065167" containerName="marketplace-operator" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395703 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="15c5f8a1-4662-457f-962a-c7ba801056e8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395717 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b35552-6680-4a88-8269-9c3e6da1ffb8" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395731 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0237c051-807d-4c6c-8f72-71cd35f2a8df" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395750 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b458ace-89d3-4a6f-950e-a7f54f90f340" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.395762 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="05309d39-9cfb-430f-b69f-9ace1978796f" containerName="registry-server" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.396761 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.400602 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.404921 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jcfpr"] Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.477643 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-utilities\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.477728 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-catalog-content\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.477758 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n858q\" (UniqueName: \"kubernetes.io/projected/eb389473-d315-4d60-830b-a0d513d4e07b-kube-api-access-n858q\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.579751 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-utilities\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.579843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-catalog-content\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.579885 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n858q\" (UniqueName: \"kubernetes.io/projected/eb389473-d315-4d60-830b-a0d513d4e07b-kube-api-access-n858q\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.580420 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-utilities\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.580686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb389473-d315-4d60-830b-a0d513d4e07b-catalog-content\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.602089 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n858q\" (UniqueName: \"kubernetes.io/projected/eb389473-d315-4d60-830b-a0d513d4e07b-kube-api-access-n858q\") pod \"certified-operators-jcfpr\" (UID: \"eb389473-d315-4d60-830b-a0d513d4e07b\") " pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.716376 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.953847 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jcfpr"] Dec 01 19:54:38 crc kubenswrapper[4960]: W1201 19:54:38.957219 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb389473_d315_4d60_830b_a0d513d4e07b.slice/crio-fc9ff863356d240ef4a0d062dc4caa9447598bc2c5255b2dee057e3e9903a035 WatchSource:0}: Error finding container fc9ff863356d240ef4a0d062dc4caa9447598bc2c5255b2dee057e3e9903a035: Status 404 returned error can't find the container with id fc9ff863356d240ef4a0d062dc4caa9447598bc2c5255b2dee057e3e9903a035 Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.992488 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k4nsz"] Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.996459 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:38 crc kubenswrapper[4960]: I1201 19:54:38.998011 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k4nsz"] Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.000283 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.000547 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerStarted","Data":"9eec3dc501ae194515b8fd09102958f32c0c54c7577d2ab688d4934d10d5c682"} Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.002074 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcfpr" event={"ID":"eb389473-d315-4d60-830b-a0d513d4e07b","Type":"ContainerStarted","Data":"fc9ff863356d240ef4a0d062dc4caa9447598bc2c5255b2dee057e3e9903a035"} Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.002341 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.006197 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qsjbd" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.085004 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-catalog-content\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.085090 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp9n5\" (UniqueName: \"kubernetes.io/projected/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-kube-api-access-bp9n5\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.085143 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-utilities\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.186817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-catalog-content\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.187129 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp9n5\" (UniqueName: \"kubernetes.io/projected/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-kube-api-access-bp9n5\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.187268 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-utilities\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.187415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-catalog-content\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.187767 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-utilities\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.211338 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp9n5\" (UniqueName: \"kubernetes.io/projected/ccd8de2a-98a4-4403-9f25-3bb5b7e8d765-kube-api-access-bp9n5\") pod \"redhat-marketplace-k4nsz\" (UID: \"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765\") " pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.363319 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:39 crc kubenswrapper[4960]: I1201 19:54:39.853861 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k4nsz"] Dec 01 19:54:39 crc kubenswrapper[4960]: W1201 19:54:39.864586 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccd8de2a_98a4_4403_9f25_3bb5b7e8d765.slice/crio-3625809a57e4386ed07db188a499097d2e216bed4089f40f58aa85b0077b56f5 WatchSource:0}: Error finding container 3625809a57e4386ed07db188a499097d2e216bed4089f40f58aa85b0077b56f5: Status 404 returned error can't find the container with id 3625809a57e4386ed07db188a499097d2e216bed4089f40f58aa85b0077b56f5 Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.015327 4960 generic.go:334] "Generic (PLEG): container finished" podID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerID="9eec3dc501ae194515b8fd09102958f32c0c54c7577d2ab688d4934d10d5c682" exitCode=0 Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.015571 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerDied","Data":"9eec3dc501ae194515b8fd09102958f32c0c54c7577d2ab688d4934d10d5c682"} Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.019379 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb389473-d315-4d60-830b-a0d513d4e07b" containerID="3052a0b8540b765a26fe35d35950dad695fc6783cccc946155db637283a4e4b7" exitCode=0 Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.019491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcfpr" event={"ID":"eb389473-d315-4d60-830b-a0d513d4e07b","Type":"ContainerDied","Data":"3052a0b8540b765a26fe35d35950dad695fc6783cccc946155db637283a4e4b7"} Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.023796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4nsz" event={"ID":"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765","Type":"ContainerStarted","Data":"3625809a57e4386ed07db188a499097d2e216bed4089f40f58aa85b0077b56f5"} Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.796012 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c4n7v"] Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.798465 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.802311 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.809622 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp88q\" (UniqueName: \"kubernetes.io/projected/81be1171-9e3d-49fd-924e-044b7d7c5023-kube-api-access-cp88q\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.809692 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-utilities\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.809775 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-catalog-content\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.814734 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4n7v"] Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.910555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp88q\" (UniqueName: \"kubernetes.io/projected/81be1171-9e3d-49fd-924e-044b7d7c5023-kube-api-access-cp88q\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.910634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-utilities\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.910708 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-catalog-content\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.911528 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-catalog-content\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.912550 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81be1171-9e3d-49fd-924e-044b7d7c5023-utilities\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:40 crc kubenswrapper[4960]: I1201 19:54:40.942315 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp88q\" (UniqueName: \"kubernetes.io/projected/81be1171-9e3d-49fd-924e-044b7d7c5023-kube-api-access-cp88q\") pod \"redhat-operators-c4n7v\" (UID: \"81be1171-9e3d-49fd-924e-044b7d7c5023\") " pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.035046 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerStarted","Data":"7138c9383ea26f4b90372a3d511bc5969aee0ca05ef00f04bc16bb5ae43e04fe"} Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.036896 4960 generic.go:334] "Generic (PLEG): container finished" podID="ccd8de2a-98a4-4403-9f25-3bb5b7e8d765" containerID="da156e97d59265e0b66d4a54935394692fe650e718f84ac7e76dd5532adc2860" exitCode=0 Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.038179 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4nsz" event={"ID":"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765","Type":"ContainerDied","Data":"da156e97d59265e0b66d4a54935394692fe650e718f84ac7e76dd5532adc2860"} Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.061792 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wcx2" podStartSLOduration=2.5017220030000002 podStartE2EDuration="5.061756266s" podCreationTimestamp="2025-12-01 19:54:36 +0000 UTC" firstStartedPulling="2025-12-01 19:54:37.979677253 +0000 UTC m=+913.267168922" lastFinishedPulling="2025-12-01 19:54:40.539711486 +0000 UTC m=+915.827203185" observedRunningTime="2025-12-01 19:54:41.057174245 +0000 UTC m=+916.344665934" watchObservedRunningTime="2025-12-01 19:54:41.061756266 +0000 UTC m=+916.349247935" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.124463 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.390404 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8v7bs"] Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.392368 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.400285 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8v7bs"] Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.416839 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-utilities\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.417006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nspcf\" (UniqueName: \"kubernetes.io/projected/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-kube-api-access-nspcf\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.417076 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-catalog-content\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.518748 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-utilities\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.518804 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nspcf\" (UniqueName: \"kubernetes.io/projected/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-kube-api-access-nspcf\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.518831 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-catalog-content\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.519326 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-catalog-content\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.519676 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-utilities\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.542932 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4n7v"] Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.542937 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nspcf\" (UniqueName: \"kubernetes.io/projected/bd9c07a1-998c-479c-8440-2d7a4b27bb2a-kube-api-access-nspcf\") pod \"community-operators-8v7bs\" (UID: \"bd9c07a1-998c-479c-8440-2d7a4b27bb2a\") " pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: W1201 19:54:41.555204 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81be1171_9e3d_49fd_924e_044b7d7c5023.slice/crio-33c34e5acd4c115aa0efb4e616227285c5e5255a898ee2778ace1315b8f994db WatchSource:0}: Error finding container 33c34e5acd4c115aa0efb4e616227285c5e5255a898ee2778ace1315b8f994db: Status 404 returned error can't find the container with id 33c34e5acd4c115aa0efb4e616227285c5e5255a898ee2778ace1315b8f994db Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.711865 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:41 crc kubenswrapper[4960]: I1201 19:54:41.940540 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8v7bs"] Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.046581 4960 generic.go:334] "Generic (PLEG): container finished" podID="eb389473-d315-4d60-830b-a0d513d4e07b" containerID="67c61aa9ce3f391ba935bf4ea631ae2706d3091e9ec14e481a53b70cadfcb5e4" exitCode=0 Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.046721 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcfpr" event={"ID":"eb389473-d315-4d60-830b-a0d513d4e07b","Type":"ContainerDied","Data":"67c61aa9ce3f391ba935bf4ea631ae2706d3091e9ec14e481a53b70cadfcb5e4"} Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.048952 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8v7bs" event={"ID":"bd9c07a1-998c-479c-8440-2d7a4b27bb2a","Type":"ContainerStarted","Data":"a031cf837663a69a06fc32859890f314aca53fe0ee4d18d5ead5bac295d7d967"} Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.050621 4960 generic.go:334] "Generic (PLEG): container finished" podID="81be1171-9e3d-49fd-924e-044b7d7c5023" containerID="4b99b6b56d71fdcbd086c417c60f6e63bf6fd66fd3bd733e5d816d4383e57726" exitCode=0 Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.050692 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4n7v" event={"ID":"81be1171-9e3d-49fd-924e-044b7d7c5023","Type":"ContainerDied","Data":"4b99b6b56d71fdcbd086c417c60f6e63bf6fd66fd3bd733e5d816d4383e57726"} Dec 01 19:54:42 crc kubenswrapper[4960]: I1201 19:54:42.050743 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4n7v" event={"ID":"81be1171-9e3d-49fd-924e-044b7d7c5023","Type":"ContainerStarted","Data":"33c34e5acd4c115aa0efb4e616227285c5e5255a898ee2778ace1315b8f994db"} Dec 01 19:54:43 crc kubenswrapper[4960]: I1201 19:54:43.059988 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcfpr" event={"ID":"eb389473-d315-4d60-830b-a0d513d4e07b","Type":"ContainerStarted","Data":"2a86a55eb80f4a410f0b4105c3157ed054d3a0f667969d3a3fbec359aa1d536c"} Dec 01 19:54:43 crc kubenswrapper[4960]: I1201 19:54:43.062273 4960 generic.go:334] "Generic (PLEG): container finished" podID="bd9c07a1-998c-479c-8440-2d7a4b27bb2a" containerID="63d0eace67b3081decaf2118a938d071ae29ab423d464b10eacb86957ee50864" exitCode=0 Dec 01 19:54:43 crc kubenswrapper[4960]: I1201 19:54:43.062321 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8v7bs" event={"ID":"bd9c07a1-998c-479c-8440-2d7a4b27bb2a","Type":"ContainerDied","Data":"63d0eace67b3081decaf2118a938d071ae29ab423d464b10eacb86957ee50864"} Dec 01 19:54:43 crc kubenswrapper[4960]: I1201 19:54:43.087740 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jcfpr" podStartSLOduration=2.593885511 podStartE2EDuration="5.087710638s" podCreationTimestamp="2025-12-01 19:54:38 +0000 UTC" firstStartedPulling="2025-12-01 19:54:40.022672752 +0000 UTC m=+915.310164451" lastFinishedPulling="2025-12-01 19:54:42.516497909 +0000 UTC m=+917.803989578" observedRunningTime="2025-12-01 19:54:43.079305068 +0000 UTC m=+918.366796757" watchObservedRunningTime="2025-12-01 19:54:43.087710638 +0000 UTC m=+918.375202307" Dec 01 19:54:44 crc kubenswrapper[4960]: I1201 19:54:44.084204 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8v7bs" event={"ID":"bd9c07a1-998c-479c-8440-2d7a4b27bb2a","Type":"ContainerStarted","Data":"6cd1b557ee2719db58947baac18f96f55de7d2f2361e42e5ac3fcbf018ee791e"} Dec 01 19:54:44 crc kubenswrapper[4960]: I1201 19:54:44.094408 4960 generic.go:334] "Generic (PLEG): container finished" podID="81be1171-9e3d-49fd-924e-044b7d7c5023" containerID="db03b2da39cff5f7047107ccf6553a5a07b1258d45a01715c9899a884be0630c" exitCode=0 Dec 01 19:54:44 crc kubenswrapper[4960]: I1201 19:54:44.094476 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4n7v" event={"ID":"81be1171-9e3d-49fd-924e-044b7d7c5023","Type":"ContainerDied","Data":"db03b2da39cff5f7047107ccf6553a5a07b1258d45a01715c9899a884be0630c"} Dec 01 19:54:45 crc kubenswrapper[4960]: I1201 19:54:45.106078 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8v7bs" event={"ID":"bd9c07a1-998c-479c-8440-2d7a4b27bb2a","Type":"ContainerDied","Data":"6cd1b557ee2719db58947baac18f96f55de7d2f2361e42e5ac3fcbf018ee791e"} Dec 01 19:54:45 crc kubenswrapper[4960]: I1201 19:54:45.105955 4960 generic.go:334] "Generic (PLEG): container finished" podID="bd9c07a1-998c-479c-8440-2d7a4b27bb2a" containerID="6cd1b557ee2719db58947baac18f96f55de7d2f2361e42e5ac3fcbf018ee791e" exitCode=0 Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.114561 4960 generic.go:334] "Generic (PLEG): container finished" podID="ccd8de2a-98a4-4403-9f25-3bb5b7e8d765" containerID="fdbb96e626311e4fe7e65027132b9bb9bbb66ac83fdb15739e6800d3b94a16d0" exitCode=0 Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.114697 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4nsz" event={"ID":"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765","Type":"ContainerDied","Data":"fdbb96e626311e4fe7e65027132b9bb9bbb66ac83fdb15739e6800d3b94a16d0"} Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.117673 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4n7v" event={"ID":"81be1171-9e3d-49fd-924e-044b7d7c5023","Type":"ContainerStarted","Data":"4933b279bf96b584e3d8c9671a5d08a6fc4e5eccf9635c6ff5418b765a74354b"} Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.120095 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8v7bs" event={"ID":"bd9c07a1-998c-479c-8440-2d7a4b27bb2a","Type":"ContainerStarted","Data":"a737138c14294ce8243822ccf08ff35c2cef161819eaa63670fd3b192338c082"} Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.156197 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8v7bs" podStartSLOduration=2.591425251 podStartE2EDuration="5.15616943s" podCreationTimestamp="2025-12-01 19:54:41 +0000 UTC" firstStartedPulling="2025-12-01 19:54:43.064963655 +0000 UTC m=+918.352455324" lastFinishedPulling="2025-12-01 19:54:45.629707834 +0000 UTC m=+920.917199503" observedRunningTime="2025-12-01 19:54:46.151010911 +0000 UTC m=+921.438502590" watchObservedRunningTime="2025-12-01 19:54:46.15616943 +0000 UTC m=+921.443661109" Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.178950 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c4n7v" podStartSLOduration=3.285670928 podStartE2EDuration="6.178931884s" podCreationTimestamp="2025-12-01 19:54:40 +0000 UTC" firstStartedPulling="2025-12-01 19:54:42.052439713 +0000 UTC m=+917.339931382" lastFinishedPulling="2025-12-01 19:54:44.945700659 +0000 UTC m=+920.233192338" observedRunningTime="2025-12-01 19:54:46.174462626 +0000 UTC m=+921.461954315" watchObservedRunningTime="2025-12-01 19:54:46.178931884 +0000 UTC m=+921.466423553" Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.989714 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:46 crc kubenswrapper[4960]: I1201 19:54:46.989781 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:47 crc kubenswrapper[4960]: I1201 19:54:47.038011 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:47 crc kubenswrapper[4960]: I1201 19:54:47.128745 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k4nsz" event={"ID":"ccd8de2a-98a4-4403-9f25-3bb5b7e8d765","Type":"ContainerStarted","Data":"be5027410e34df672b910271b331d7bd1dbb9951b841e8a28d31b903f85936c5"} Dec 01 19:54:47 crc kubenswrapper[4960]: I1201 19:54:47.155622 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k4nsz" podStartSLOduration=3.56099898 podStartE2EDuration="9.155601908s" podCreationTimestamp="2025-12-01 19:54:38 +0000 UTC" firstStartedPulling="2025-12-01 19:54:41.039380544 +0000 UTC m=+916.326872213" lastFinishedPulling="2025-12-01 19:54:46.633983472 +0000 UTC m=+921.921475141" observedRunningTime="2025-12-01 19:54:47.15210129 +0000 UTC m=+922.439592969" watchObservedRunningTime="2025-12-01 19:54:47.155601908 +0000 UTC m=+922.443093577" Dec 01 19:54:47 crc kubenswrapper[4960]: I1201 19:54:47.174032 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:48 crc kubenswrapper[4960]: I1201 19:54:48.717078 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:48 crc kubenswrapper[4960]: I1201 19:54:48.717186 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:48 crc kubenswrapper[4960]: I1201 19:54:48.772414 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:49 crc kubenswrapper[4960]: I1201 19:54:49.187775 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jcfpr" Dec 01 19:54:49 crc kubenswrapper[4960]: I1201 19:54:49.364485 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:49 crc kubenswrapper[4960]: I1201 19:54:49.364559 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:49 crc kubenswrapper[4960]: I1201 19:54:49.409081 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:54:49 crc kubenswrapper[4960]: I1201 19:54:49.784369 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:50 crc kubenswrapper[4960]: I1201 19:54:50.150197 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wcx2" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="registry-server" containerID="cri-o://7138c9383ea26f4b90372a3d511bc5969aee0ca05ef00f04bc16bb5ae43e04fe" gracePeriod=2 Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.125400 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.125602 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.178877 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.713034 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.713918 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:51 crc kubenswrapper[4960]: I1201 19:54:51.786377 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.169099 4960 generic.go:334] "Generic (PLEG): container finished" podID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerID="7138c9383ea26f4b90372a3d511bc5969aee0ca05ef00f04bc16bb5ae43e04fe" exitCode=0 Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.169201 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerDied","Data":"7138c9383ea26f4b90372a3d511bc5969aee0ca05ef00f04bc16bb5ae43e04fe"} Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.219912 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8v7bs" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.240582 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c4n7v" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.552884 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.582678 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities\") pod \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.582813 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r22d\" (UniqueName: \"kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d\") pod \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.582870 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content\") pod \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\" (UID: \"b71eafc4-b189-4d2f-ad28-be0b1024d1c9\") " Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.584936 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities" (OuterVolumeSpecName: "utilities") pod "b71eafc4-b189-4d2f-ad28-be0b1024d1c9" (UID: "b71eafc4-b189-4d2f-ad28-be0b1024d1c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.595601 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d" (OuterVolumeSpecName: "kube-api-access-8r22d") pod "b71eafc4-b189-4d2f-ad28-be0b1024d1c9" (UID: "b71eafc4-b189-4d2f-ad28-be0b1024d1c9"). InnerVolumeSpecName "kube-api-access-8r22d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.647317 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b71eafc4-b189-4d2f-ad28-be0b1024d1c9" (UID: "b71eafc4-b189-4d2f-ad28-be0b1024d1c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.684936 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r22d\" (UniqueName: \"kubernetes.io/projected/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-kube-api-access-8r22d\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.684998 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:52 crc kubenswrapper[4960]: I1201 19:54:52.685016 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b71eafc4-b189-4d2f-ad28-be0b1024d1c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.177464 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wcx2" Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.185622 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wcx2" event={"ID":"b71eafc4-b189-4d2f-ad28-be0b1024d1c9","Type":"ContainerDied","Data":"a2f167d3ae66a84e84da93254eec4603bccd210b2c69ed677110c83b5a2b9d18"} Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.185672 4960 scope.go:117] "RemoveContainer" containerID="7138c9383ea26f4b90372a3d511bc5969aee0ca05ef00f04bc16bb5ae43e04fe" Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.212717 4960 scope.go:117] "RemoveContainer" containerID="9eec3dc501ae194515b8fd09102958f32c0c54c7577d2ab688d4934d10d5c682" Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.215219 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.219352 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wcx2"] Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.236615 4960 scope.go:117] "RemoveContainer" containerID="94f1cf8a1915f28ee6149f17ad6448760a5bc6a827ed7dab9c0924255239fa72" Dec 01 19:54:53 crc kubenswrapper[4960]: I1201 19:54:53.337074 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" path="/var/lib/kubelet/pods/b71eafc4-b189-4d2f-ad28-be0b1024d1c9/volumes" Dec 01 19:54:59 crc kubenswrapper[4960]: I1201 19:54:59.406375 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k4nsz" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.645890 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n"] Dec 01 19:55:01 crc kubenswrapper[4960]: E1201 19:55:01.646865 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="registry-server" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.646889 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="registry-server" Dec 01 19:55:01 crc kubenswrapper[4960]: E1201 19:55:01.646906 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="extract-content" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.646918 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="extract-content" Dec 01 19:55:01 crc kubenswrapper[4960]: E1201 19:55:01.646937 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="extract-utilities" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.646949 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="extract-utilities" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.647199 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71eafc4-b189-4d2f-ad28-be0b1024d1c9" containerName="registry-server" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.648663 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.651754 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.655184 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n"] Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.805642 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.805753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk7jq\" (UniqueName: \"kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.805807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.907240 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk7jq\" (UniqueName: \"kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.907337 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.907396 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.907937 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.908031 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.947467 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk7jq\" (UniqueName: \"kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:01 crc kubenswrapper[4960]: I1201 19:55:01.971632 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:02 crc kubenswrapper[4960]: I1201 19:55:02.168652 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n"] Dec 01 19:55:02 crc kubenswrapper[4960]: I1201 19:55:02.235835 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" event={"ID":"2ff9d002-1e21-49cf-bced-1686ee42fe90","Type":"ContainerStarted","Data":"b24e4b49ad3283794f51c91975c03e527117140f131036b09426ead6d4b8da88"} Dec 01 19:55:03 crc kubenswrapper[4960]: I1201 19:55:03.244133 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerID="9b5edfce3c03753383172637a924e60e7b0a025a7b6920667500fe756001a6fa" exitCode=0 Dec 01 19:55:03 crc kubenswrapper[4960]: I1201 19:55:03.244192 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" event={"ID":"2ff9d002-1e21-49cf-bced-1686ee42fe90","Type":"ContainerDied","Data":"9b5edfce3c03753383172637a924e60e7b0a025a7b6920667500fe756001a6fa"} Dec 01 19:55:05 crc kubenswrapper[4960]: I1201 19:55:05.258929 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerID="154e046af0ead75d2de5d95403854097ec9ebc7c151c12ab0668a2ab6e0e75c1" exitCode=0 Dec 01 19:55:05 crc kubenswrapper[4960]: I1201 19:55:05.258992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" event={"ID":"2ff9d002-1e21-49cf-bced-1686ee42fe90","Type":"ContainerDied","Data":"154e046af0ead75d2de5d95403854097ec9ebc7c151c12ab0668a2ab6e0e75c1"} Dec 01 19:55:06 crc kubenswrapper[4960]: I1201 19:55:06.278072 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerID="51f57f9bd612031097d1e2068d0ce7108a54efb8fe2b2121b6adebf224f98f16" exitCode=0 Dec 01 19:55:06 crc kubenswrapper[4960]: I1201 19:55:06.278159 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" event={"ID":"2ff9d002-1e21-49cf-bced-1686ee42fe90","Type":"ContainerDied","Data":"51f57f9bd612031097d1e2068d0ce7108a54efb8fe2b2121b6adebf224f98f16"} Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.572931 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.684232 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util\") pod \"2ff9d002-1e21-49cf-bced-1686ee42fe90\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.684333 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle\") pod \"2ff9d002-1e21-49cf-bced-1686ee42fe90\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.684383 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk7jq\" (UniqueName: \"kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq\") pod \"2ff9d002-1e21-49cf-bced-1686ee42fe90\" (UID: \"2ff9d002-1e21-49cf-bced-1686ee42fe90\") " Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.685321 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle" (OuterVolumeSpecName: "bundle") pod "2ff9d002-1e21-49cf-bced-1686ee42fe90" (UID: "2ff9d002-1e21-49cf-bced-1686ee42fe90"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.691908 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq" (OuterVolumeSpecName: "kube-api-access-rk7jq") pod "2ff9d002-1e21-49cf-bced-1686ee42fe90" (UID: "2ff9d002-1e21-49cf-bced-1686ee42fe90"). InnerVolumeSpecName "kube-api-access-rk7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.701647 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util" (OuterVolumeSpecName: "util") pod "2ff9d002-1e21-49cf-bced-1686ee42fe90" (UID: "2ff9d002-1e21-49cf-bced-1686ee42fe90"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.785895 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.785939 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk7jq\" (UniqueName: \"kubernetes.io/projected/2ff9d002-1e21-49cf-bced-1686ee42fe90-kube-api-access-rk7jq\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:07 crc kubenswrapper[4960]: I1201 19:55:07.785951 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2ff9d002-1e21-49cf-bced-1686ee42fe90-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:08 crc kubenswrapper[4960]: I1201 19:55:08.293748 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" event={"ID":"2ff9d002-1e21-49cf-bced-1686ee42fe90","Type":"ContainerDied","Data":"b24e4b49ad3283794f51c91975c03e527117140f131036b09426ead6d4b8da88"} Dec 01 19:55:08 crc kubenswrapper[4960]: I1201 19:55:08.293797 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b24e4b49ad3283794f51c91975c03e527117140f131036b09426ead6d4b8da88" Dec 01 19:55:08 crc kubenswrapper[4960]: I1201 19:55:08.293829 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.849847 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd"] Dec 01 19:55:13 crc kubenswrapper[4960]: E1201 19:55:13.850787 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="extract" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.850812 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="extract" Dec 01 19:55:13 crc kubenswrapper[4960]: E1201 19:55:13.850837 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="pull" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.850849 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="pull" Dec 01 19:55:13 crc kubenswrapper[4960]: E1201 19:55:13.850881 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="util" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.850893 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="util" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.851091 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff9d002-1e21-49cf-bced-1686ee42fe90" containerName="extract" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.851905 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.856522 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.857617 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-qk4zs" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.857722 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.865208 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd"] Dec 01 19:55:13 crc kubenswrapper[4960]: I1201 19:55:13.967602 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f8qc\" (UniqueName: \"kubernetes.io/projected/6377bee0-949f-4c1b-9758-c425ab4660bb-kube-api-access-5f8qc\") pod \"nmstate-operator-5b5b58f5c8-bvgrd\" (UID: \"6377bee0-949f-4c1b-9758-c425ab4660bb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" Dec 01 19:55:14 crc kubenswrapper[4960]: I1201 19:55:14.068937 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f8qc\" (UniqueName: \"kubernetes.io/projected/6377bee0-949f-4c1b-9758-c425ab4660bb-kube-api-access-5f8qc\") pod \"nmstate-operator-5b5b58f5c8-bvgrd\" (UID: \"6377bee0-949f-4c1b-9758-c425ab4660bb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" Dec 01 19:55:14 crc kubenswrapper[4960]: I1201 19:55:14.107193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f8qc\" (UniqueName: \"kubernetes.io/projected/6377bee0-949f-4c1b-9758-c425ab4660bb-kube-api-access-5f8qc\") pod \"nmstate-operator-5b5b58f5c8-bvgrd\" (UID: \"6377bee0-949f-4c1b-9758-c425ab4660bb\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" Dec 01 19:55:14 crc kubenswrapper[4960]: I1201 19:55:14.189806 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" Dec 01 19:55:14 crc kubenswrapper[4960]: I1201 19:55:14.528593 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd"] Dec 01 19:55:15 crc kubenswrapper[4960]: I1201 19:55:15.350491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" event={"ID":"6377bee0-949f-4c1b-9758-c425ab4660bb","Type":"ContainerStarted","Data":"8ec7a7f7dd2044f02caac63ac18ed761317735f981713fd7a0856347b7107ff0"} Dec 01 19:55:17 crc kubenswrapper[4960]: I1201 19:55:17.365342 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" event={"ID":"6377bee0-949f-4c1b-9758-c425ab4660bb","Type":"ContainerStarted","Data":"3c04fd7d9d52adc53119f2c6160ae45a08afb197127cb663b8a79beaabc1d62f"} Dec 01 19:55:17 crc kubenswrapper[4960]: I1201 19:55:17.392165 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-bvgrd" podStartSLOduration=2.448956593 podStartE2EDuration="4.392136567s" podCreationTimestamp="2025-12-01 19:55:13 +0000 UTC" firstStartedPulling="2025-12-01 19:55:14.539458876 +0000 UTC m=+949.826950545" lastFinishedPulling="2025-12-01 19:55:16.48263885 +0000 UTC m=+951.770130519" observedRunningTime="2025-12-01 19:55:17.38898948 +0000 UTC m=+952.676481159" watchObservedRunningTime="2025-12-01 19:55:17.392136567 +0000 UTC m=+952.679628276" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.572300 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.573860 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.576255 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xvwmk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.596009 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.603439 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.604424 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.607431 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.637097 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.642595 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-bwb2n"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.643445 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.700480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/138d040c-3866-4963-a2c8-26fb8a6c9811-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.700537 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4hkl\" (UniqueName: \"kubernetes.io/projected/138d040c-3866-4963-a2c8-26fb8a6c9811-kube-api-access-x4hkl\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.700576 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6sk\" (UniqueName: \"kubernetes.io/projected/cba17f81-92f7-418b-a5c7-a184634c9acc-kube-api-access-bk6sk\") pod \"nmstate-metrics-7f946cbc9-vh2zk\" (UID: \"cba17f81-92f7-418b-a5c7-a184634c9acc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.752979 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.753748 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.755851 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.756042 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-fqh4f" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.762649 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.765799 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.801791 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9pns\" (UniqueName: \"kubernetes.io/projected/80d93577-3e89-47d7-b363-f22021509688-kube-api-access-f9pns\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.801865 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-dbus-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.801992 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/138d040c-3866-4963-a2c8-26fb8a6c9811-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.802041 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4hkl\" (UniqueName: \"kubernetes.io/projected/138d040c-3866-4963-a2c8-26fb8a6c9811-kube-api-access-x4hkl\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.802785 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-nmstate-lock\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.802876 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6sk\" (UniqueName: \"kubernetes.io/projected/cba17f81-92f7-418b-a5c7-a184634c9acc-kube-api-access-bk6sk\") pod \"nmstate-metrics-7f946cbc9-vh2zk\" (UID: \"cba17f81-92f7-418b-a5c7-a184634c9acc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.802921 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-ovs-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.809325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/138d040c-3866-4963-a2c8-26fb8a6c9811-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.820711 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6sk\" (UniqueName: \"kubernetes.io/projected/cba17f81-92f7-418b-a5c7-a184634c9acc-kube-api-access-bk6sk\") pod \"nmstate-metrics-7f946cbc9-vh2zk\" (UID: \"cba17f81-92f7-418b-a5c7-a184634c9acc\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.821876 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4hkl\" (UniqueName: \"kubernetes.io/projected/138d040c-3866-4963-a2c8-26fb8a6c9811-kube-api-access-x4hkl\") pod \"nmstate-webhook-5f6d4c5ccb-5ndvb\" (UID: \"138d040c-3866-4963-a2c8-26fb8a6c9811\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.890840 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.904774 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9pns\" (UniqueName: \"kubernetes.io/projected/80d93577-3e89-47d7-b363-f22021509688-kube-api-access-f9pns\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.904858 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-dbus-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.904928 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-nmstate-lock\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.904952 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/751bcd73-4c55-4fbb-a12d-0c9956a63e16-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905036 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/751bcd73-4c55-4fbb-a12d-0c9956a63e16-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905092 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-ovs-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905081 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-nmstate-lock\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905220 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-ovs-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905258 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6dd2\" (UniqueName: \"kubernetes.io/projected/751bcd73-4c55-4fbb-a12d-0c9956a63e16-kube-api-access-r6dd2\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.905309 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/80d93577-3e89-47d7-b363-f22021509688-dbus-socket\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.918527 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.929710 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9pns\" (UniqueName: \"kubernetes.io/projected/80d93577-3e89-47d7-b363-f22021509688-kube-api-access-f9pns\") pod \"nmstate-handler-bwb2n\" (UID: \"80d93577-3e89-47d7-b363-f22021509688\") " pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.945065 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-764c8dccb-xml97"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.950686 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.954241 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-764c8dccb-xml97"] Dec 01 19:55:22 crc kubenswrapper[4960]: I1201 19:55:22.973168 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.006277 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6dd2\" (UniqueName: \"kubernetes.io/projected/751bcd73-4c55-4fbb-a12d-0c9956a63e16-kube-api-access-r6dd2\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.006358 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/751bcd73-4c55-4fbb-a12d-0c9956a63e16-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.006385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/751bcd73-4c55-4fbb-a12d-0c9956a63e16-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.007199 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/751bcd73-4c55-4fbb-a12d-0c9956a63e16-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.012840 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/751bcd73-4c55-4fbb-a12d-0c9956a63e16-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.029325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6dd2\" (UniqueName: \"kubernetes.io/projected/751bcd73-4c55-4fbb-a12d-0c9956a63e16-kube-api-access-r6dd2\") pod \"nmstate-console-plugin-7fbb5f6569-xpq6m\" (UID: \"751bcd73-4c55-4fbb-a12d-0c9956a63e16\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.075865 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108303 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-oauth-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-978nj\" (UniqueName: \"kubernetes.io/projected/2781ec3d-a614-445a-8581-5e11523a5fee-kube-api-access-978nj\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-service-ca\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108413 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108437 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-trusted-ca-bundle\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108457 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-console-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.108497 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-oauth-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.187658 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb"] Dec 01 19:55:23 crc kubenswrapper[4960]: W1201 19:55:23.194603 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod138d040c_3866_4963_a2c8_26fb8a6c9811.slice/crio-a9ae13f4b5e8eadbede4c72094e0bf1bacbf0b7979b080826c484b11088dedd9 WatchSource:0}: Error finding container a9ae13f4b5e8eadbede4c72094e0bf1bacbf0b7979b080826c484b11088dedd9: Status 404 returned error can't find the container with id a9ae13f4b5e8eadbede4c72094e0bf1bacbf0b7979b080826c484b11088dedd9 Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.209742 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-console-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.209843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-oauth-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.209897 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-oauth-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.209935 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-978nj\" (UniqueName: \"kubernetes.io/projected/2781ec3d-a614-445a-8581-5e11523a5fee-kube-api-access-978nj\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.210011 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-service-ca\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.210039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.210072 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-trusted-ca-bundle\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.212164 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-oauth-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.212211 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-console-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.212384 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-service-ca\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.212527 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2781ec3d-a614-445a-8581-5e11523a5fee-trusted-ca-bundle\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.216515 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-serving-cert\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.218891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2781ec3d-a614-445a-8581-5e11523a5fee-console-oauth-config\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.230708 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-978nj\" (UniqueName: \"kubernetes.io/projected/2781ec3d-a614-445a-8581-5e11523a5fee-kube-api-access-978nj\") pod \"console-764c8dccb-xml97\" (UID: \"2781ec3d-a614-445a-8581-5e11523a5fee\") " pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.294997 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.349394 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk"] Dec 01 19:55:23 crc kubenswrapper[4960]: W1201 19:55:23.362402 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcba17f81_92f7_418b_a5c7_a184634c9acc.slice/crio-75d914616aefd5f56b882bfe5fde4512c7f4beb18b48dd67757320d122072d59 WatchSource:0}: Error finding container 75d914616aefd5f56b882bfe5fde4512c7f4beb18b48dd67757320d122072d59: Status 404 returned error can't find the container with id 75d914616aefd5f56b882bfe5fde4512c7f4beb18b48dd67757320d122072d59 Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.421251 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bwb2n" event={"ID":"80d93577-3e89-47d7-b363-f22021509688","Type":"ContainerStarted","Data":"fcaa4087a5e842735197d5f5a8912e105d8e2c7ad5bacb1e4ce27843fb54da90"} Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.422654 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" event={"ID":"138d040c-3866-4963-a2c8-26fb8a6c9811","Type":"ContainerStarted","Data":"a9ae13f4b5e8eadbede4c72094e0bf1bacbf0b7979b080826c484b11088dedd9"} Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.424752 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" event={"ID":"cba17f81-92f7-418b-a5c7-a184634c9acc","Type":"ContainerStarted","Data":"75d914616aefd5f56b882bfe5fde4512c7f4beb18b48dd67757320d122072d59"} Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.488408 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m"] Dec 01 19:55:23 crc kubenswrapper[4960]: I1201 19:55:23.521072 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-764c8dccb-xml97"] Dec 01 19:55:23 crc kubenswrapper[4960]: W1201 19:55:23.523923 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2781ec3d_a614_445a_8581_5e11523a5fee.slice/crio-70b204b0a5f6e00eb9eb103e87e3f9fd04e09991bd4f14c25ffbe8211cc8696f WatchSource:0}: Error finding container 70b204b0a5f6e00eb9eb103e87e3f9fd04e09991bd4f14c25ffbe8211cc8696f: Status 404 returned error can't find the container with id 70b204b0a5f6e00eb9eb103e87e3f9fd04e09991bd4f14c25ffbe8211cc8696f Dec 01 19:55:24 crc kubenswrapper[4960]: I1201 19:55:24.439535 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-764c8dccb-xml97" event={"ID":"2781ec3d-a614-445a-8581-5e11523a5fee","Type":"ContainerStarted","Data":"c34b281c4f1903a729b8e620c08f248d2ad9f9a9c496c69b225fc2038b5c7185"} Dec 01 19:55:24 crc kubenswrapper[4960]: I1201 19:55:24.440152 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-764c8dccb-xml97" event={"ID":"2781ec3d-a614-445a-8581-5e11523a5fee","Type":"ContainerStarted","Data":"70b204b0a5f6e00eb9eb103e87e3f9fd04e09991bd4f14c25ffbe8211cc8696f"} Dec 01 19:55:24 crc kubenswrapper[4960]: I1201 19:55:24.441934 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" event={"ID":"751bcd73-4c55-4fbb-a12d-0c9956a63e16","Type":"ContainerStarted","Data":"22b220b49a369577c1557b02b9f3dc185093295e5d1ce928a964143bee1134e3"} Dec 01 19:55:24 crc kubenswrapper[4960]: I1201 19:55:24.465728 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-764c8dccb-xml97" podStartSLOduration=2.465706939 podStartE2EDuration="2.465706939s" podCreationTimestamp="2025-12-01 19:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:55:24.460080355 +0000 UTC m=+959.747572044" watchObservedRunningTime="2025-12-01 19:55:24.465706939 +0000 UTC m=+959.753198608" Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.458333 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" event={"ID":"138d040c-3866-4963-a2c8-26fb8a6c9811","Type":"ContainerStarted","Data":"f1094e54ad9959cfc5c351e44d64676650b0a59ae8a83a5d33f19f0f00db915a"} Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.461302 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.461389 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" event={"ID":"cba17f81-92f7-418b-a5c7-a184634c9acc","Type":"ContainerStarted","Data":"8373639b6eb092cfb3660e1fa086eb9bfcdfb1aadc9e136a2717d97d107a4885"} Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.463446 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-bwb2n" event={"ID":"80d93577-3e89-47d7-b363-f22021509688","Type":"ContainerStarted","Data":"1ffe851317e005bfe69ae7cb961e7ea7f7df82a2d0357e8e012cefd7bbbbcc6a"} Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.463783 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.478518 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" podStartSLOduration=2.041749703 podStartE2EDuration="4.478489585s" podCreationTimestamp="2025-12-01 19:55:22 +0000 UTC" firstStartedPulling="2025-12-01 19:55:23.198689919 +0000 UTC m=+958.486181588" lastFinishedPulling="2025-12-01 19:55:25.635429801 +0000 UTC m=+960.922921470" observedRunningTime="2025-12-01 19:55:26.474001496 +0000 UTC m=+961.761493175" watchObservedRunningTime="2025-12-01 19:55:26.478489585 +0000 UTC m=+961.765981254" Dec 01 19:55:26 crc kubenswrapper[4960]: I1201 19:55:26.503723 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-bwb2n" podStartSLOduration=1.915375654 podStartE2EDuration="4.50358765s" podCreationTimestamp="2025-12-01 19:55:22 +0000 UTC" firstStartedPulling="2025-12-01 19:55:23.006958731 +0000 UTC m=+958.294450400" lastFinishedPulling="2025-12-01 19:55:25.595170727 +0000 UTC m=+960.882662396" observedRunningTime="2025-12-01 19:55:26.497638276 +0000 UTC m=+961.785129955" watchObservedRunningTime="2025-12-01 19:55:26.50358765 +0000 UTC m=+961.791079329" Dec 01 19:55:27 crc kubenswrapper[4960]: I1201 19:55:27.473492 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" event={"ID":"751bcd73-4c55-4fbb-a12d-0c9956a63e16","Type":"ContainerStarted","Data":"e7f4bec377b369422de83502c793ff37d29cd15959556fafeb6877c89ef00427"} Dec 01 19:55:27 crc kubenswrapper[4960]: I1201 19:55:27.498272 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xpq6m" podStartSLOduration=2.299830171 podStartE2EDuration="5.498243201s" podCreationTimestamp="2025-12-01 19:55:22 +0000 UTC" firstStartedPulling="2025-12-01 19:55:23.498945661 +0000 UTC m=+958.786437350" lastFinishedPulling="2025-12-01 19:55:26.697358701 +0000 UTC m=+961.984850380" observedRunningTime="2025-12-01 19:55:27.492971298 +0000 UTC m=+962.780462977" watchObservedRunningTime="2025-12-01 19:55:27.498243201 +0000 UTC m=+962.785734880" Dec 01 19:55:29 crc kubenswrapper[4960]: I1201 19:55:29.493400 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" event={"ID":"cba17f81-92f7-418b-a5c7-a184634c9acc","Type":"ContainerStarted","Data":"e8c9e53f049b56a315b5c2883932b302d2d8092e98324bf0d743432c73a58b16"} Dec 01 19:55:29 crc kubenswrapper[4960]: I1201 19:55:29.525394 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-vh2zk" podStartSLOduration=2.459524828 podStartE2EDuration="7.525364489s" podCreationTimestamp="2025-12-01 19:55:22 +0000 UTC" firstStartedPulling="2025-12-01 19:55:23.369340655 +0000 UTC m=+958.656832314" lastFinishedPulling="2025-12-01 19:55:28.435180296 +0000 UTC m=+963.722671975" observedRunningTime="2025-12-01 19:55:29.518469756 +0000 UTC m=+964.805961505" watchObservedRunningTime="2025-12-01 19:55:29.525364489 +0000 UTC m=+964.812856198" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.015753 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-bwb2n" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.295501 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.295583 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.305516 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.537373 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-764c8dccb-xml97" Dec 01 19:55:33 crc kubenswrapper[4960]: I1201 19:55:33.611581 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:55:42 crc kubenswrapper[4960]: I1201 19:55:42.926604 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5ndvb" Dec 01 19:55:58 crc kubenswrapper[4960]: I1201 19:55:58.679819 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-j9vkv" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerName="console" containerID="cri-o://398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0" gracePeriod=15 Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.084387 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-j9vkv_ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0/console/0.log" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.084805 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214537 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214653 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214730 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214777 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214877 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb97s\" (UniqueName: \"kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.214995 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config\") pod \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\" (UID: \"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0\") " Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.216245 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.216290 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca" (OuterVolumeSpecName: "service-ca") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.216318 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config" (OuterVolumeSpecName: "console-config") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.216735 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.223503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s" (OuterVolumeSpecName: "kube-api-access-rb97s") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "kube-api-access-rb97s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.223757 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.224037 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" (UID: "ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.317468 4960 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.317534 4960 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.317553 4960 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.318165 4960 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.318219 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb97s\" (UniqueName: \"kubernetes.io/projected/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-kube-api-access-rb97s\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.318238 4960 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.318251 4960 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.761141 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-j9vkv_ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0/console/0.log" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.761873 4960 generic.go:334] "Generic (PLEG): container finished" podID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerID="398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0" exitCode=2 Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.761940 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j9vkv" event={"ID":"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0","Type":"ContainerDied","Data":"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0"} Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.761974 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-j9vkv" event={"ID":"ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0","Type":"ContainerDied","Data":"f7aa4d6f4a853d744095b449ea12619f9780956d202316f2abd8fac1e4a697df"} Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.761997 4960 scope.go:117] "RemoveContainer" containerID="398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.762373 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-j9vkv" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.781449 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.784878 4960 scope.go:117] "RemoveContainer" containerID="398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0" Dec 01 19:55:59 crc kubenswrapper[4960]: E1201 19:55:59.785384 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0\": container with ID starting with 398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0 not found: ID does not exist" containerID="398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.785435 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0"} err="failed to get container status \"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0\": rpc error: code = NotFound desc = could not find container \"398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0\": container with ID starting with 398761c835f2ca8764b580117503c2b1f4000a52a113e1dfd9cfb6c71a93d5b0 not found: ID does not exist" Dec 01 19:55:59 crc kubenswrapper[4960]: I1201 19:55:59.785464 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-j9vkv"] Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.493404 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln"] Dec 01 19:56:00 crc kubenswrapper[4960]: E1201 19:56:00.493770 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerName="console" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.493787 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerName="console" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.493928 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" containerName="console" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.494944 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.496921 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.505835 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln"] Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.535997 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnx98\" (UniqueName: \"kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.536099 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.536151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.637432 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnx98\" (UniqueName: \"kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.637561 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.637616 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.638302 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.638518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.662263 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnx98\" (UniqueName: \"kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:00 crc kubenswrapper[4960]: I1201 19:56:00.812555 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:01 crc kubenswrapper[4960]: I1201 19:56:01.048713 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln"] Dec 01 19:56:01 crc kubenswrapper[4960]: I1201 19:56:01.332876 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0" path="/var/lib/kubelet/pods/ee9aff31-fb6f-41a7-8cf1-3f6b549ec4a0/volumes" Dec 01 19:56:01 crc kubenswrapper[4960]: I1201 19:56:01.784221 4960 generic.go:334] "Generic (PLEG): container finished" podID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerID="55f6ffe1e717be21fda8546fd81e434da2ab64e3b36a0368d82797cb31f36b94" exitCode=0 Dec 01 19:56:01 crc kubenswrapper[4960]: I1201 19:56:01.784313 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" event={"ID":"96e5ebac-e485-4307-a5ff-73e77e788e02","Type":"ContainerDied","Data":"55f6ffe1e717be21fda8546fd81e434da2ab64e3b36a0368d82797cb31f36b94"} Dec 01 19:56:01 crc kubenswrapper[4960]: I1201 19:56:01.784403 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" event={"ID":"96e5ebac-e485-4307-a5ff-73e77e788e02","Type":"ContainerStarted","Data":"fb7be5e438b3c69c86eb34bc0245b4e2a93da0d47cfc19e7ec8181a7ccf87b1d"} Dec 01 19:56:03 crc kubenswrapper[4960]: I1201 19:56:03.803695 4960 generic.go:334] "Generic (PLEG): container finished" podID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerID="84cdf94ee3ddd6d2c851e8e56195bf1b19c10ee9030edfe1ee8dcb13b116d89e" exitCode=0 Dec 01 19:56:03 crc kubenswrapper[4960]: I1201 19:56:03.803785 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" event={"ID":"96e5ebac-e485-4307-a5ff-73e77e788e02","Type":"ContainerDied","Data":"84cdf94ee3ddd6d2c851e8e56195bf1b19c10ee9030edfe1ee8dcb13b116d89e"} Dec 01 19:56:04 crc kubenswrapper[4960]: I1201 19:56:04.817291 4960 generic.go:334] "Generic (PLEG): container finished" podID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerID="cbc9a53cdb523ccd9fa7191c88770868c7801986651e4e552bb026b651e4ecbe" exitCode=0 Dec 01 19:56:04 crc kubenswrapper[4960]: I1201 19:56:04.817401 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" event={"ID":"96e5ebac-e485-4307-a5ff-73e77e788e02","Type":"ContainerDied","Data":"cbc9a53cdb523ccd9fa7191c88770868c7801986651e4e552bb026b651e4ecbe"} Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.074054 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.221643 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util\") pod \"96e5ebac-e485-4307-a5ff-73e77e788e02\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.221720 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle\") pod \"96e5ebac-e485-4307-a5ff-73e77e788e02\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.222678 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnx98\" (UniqueName: \"kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98\") pod \"96e5ebac-e485-4307-a5ff-73e77e788e02\" (UID: \"96e5ebac-e485-4307-a5ff-73e77e788e02\") " Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.223044 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle" (OuterVolumeSpecName: "bundle") pod "96e5ebac-e485-4307-a5ff-73e77e788e02" (UID: "96e5ebac-e485-4307-a5ff-73e77e788e02"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.229255 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98" (OuterVolumeSpecName: "kube-api-access-dnx98") pod "96e5ebac-e485-4307-a5ff-73e77e788e02" (UID: "96e5ebac-e485-4307-a5ff-73e77e788e02"). InnerVolumeSpecName "kube-api-access-dnx98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.252059 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util" (OuterVolumeSpecName: "util") pod "96e5ebac-e485-4307-a5ff-73e77e788e02" (UID: "96e5ebac-e485-4307-a5ff-73e77e788e02"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.324179 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnx98\" (UniqueName: \"kubernetes.io/projected/96e5ebac-e485-4307-a5ff-73e77e788e02-kube-api-access-dnx98\") on node \"crc\" DevicePath \"\"" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.324251 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.324280 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/96e5ebac-e485-4307-a5ff-73e77e788e02-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.834855 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" event={"ID":"96e5ebac-e485-4307-a5ff-73e77e788e02","Type":"ContainerDied","Data":"fb7be5e438b3c69c86eb34bc0245b4e2a93da0d47cfc19e7ec8181a7ccf87b1d"} Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.834911 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln" Dec 01 19:56:06 crc kubenswrapper[4960]: I1201 19:56:06.834934 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb7be5e438b3c69c86eb34bc0245b4e2a93da0d47cfc19e7ec8181a7ccf87b1d" Dec 01 19:56:13 crc kubenswrapper[4960]: I1201 19:56:13.870912 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:56:13 crc kubenswrapper[4960]: I1201 19:56:13.871831 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.034495 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f48564cc-pv444"] Dec 01 19:56:18 crc kubenswrapper[4960]: E1201 19:56:18.034758 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="extract" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.034771 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="extract" Dec 01 19:56:18 crc kubenswrapper[4960]: E1201 19:56:18.034786 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="util" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.034792 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="util" Dec 01 19:56:18 crc kubenswrapper[4960]: E1201 19:56:18.034808 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="pull" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.034814 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="pull" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.034914 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e5ebac-e485-4307-a5ff-73e77e788e02" containerName="extract" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.035460 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.037572 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-bwprs" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.037858 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.037858 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.037856 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.042708 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.053501 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f48564cc-pv444"] Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.200423 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb"] Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.201474 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204458 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-apiservice-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204496 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-webhook-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4mb\" (UniqueName: \"kubernetes.io/projected/8b72f468-b777-407c-8e36-30cc398644a6-kube-api-access-gb4mb\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204545 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-webhook-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204569 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-apiservice-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.204595 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nrdp\" (UniqueName: \"kubernetes.io/projected/914a4b4c-d013-45ef-832d-35b600874887-kube-api-access-4nrdp\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.205590 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.205975 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6nk9q" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.206191 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.271632 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb"] Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.305781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-apiservice-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.305842 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nrdp\" (UniqueName: \"kubernetes.io/projected/914a4b4c-d013-45ef-832d-35b600874887-kube-api-access-4nrdp\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.305920 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-apiservice-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.305952 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-webhook-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.305979 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4mb\" (UniqueName: \"kubernetes.io/projected/8b72f468-b777-407c-8e36-30cc398644a6-kube-api-access-gb4mb\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.306016 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-webhook-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.312802 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-apiservice-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.312802 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-webhook-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.312830 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8b72f468-b777-407c-8e36-30cc398644a6-apiservice-cert\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.313536 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/914a4b4c-d013-45ef-832d-35b600874887-webhook-cert\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.328402 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nrdp\" (UniqueName: \"kubernetes.io/projected/914a4b4c-d013-45ef-832d-35b600874887-kube-api-access-4nrdp\") pod \"metallb-operator-webhook-server-7d7c4c664-7fttb\" (UID: \"914a4b4c-d013-45ef-832d-35b600874887\") " pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.330136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4mb\" (UniqueName: \"kubernetes.io/projected/8b72f468-b777-407c-8e36-30cc398644a6-kube-api-access-gb4mb\") pod \"metallb-operator-controller-manager-64f48564cc-pv444\" (UID: \"8b72f468-b777-407c-8e36-30cc398644a6\") " pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.354825 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.516548 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.584031 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-64f48564cc-pv444"] Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.923727 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" event={"ID":"8b72f468-b777-407c-8e36-30cc398644a6","Type":"ContainerStarted","Data":"603c6f0010f70cabc29ff9f126ad7636b9795fd18e0d63493030167969c49238"} Dec 01 19:56:18 crc kubenswrapper[4960]: I1201 19:56:18.955310 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb"] Dec 01 19:56:19 crc kubenswrapper[4960]: I1201 19:56:19.932366 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" event={"ID":"914a4b4c-d013-45ef-832d-35b600874887","Type":"ContainerStarted","Data":"898f3b2349ea88ee0b2d15ab85d5920c820bb9441a19d4d196a2748e6bceef9f"} Dec 01 19:56:23 crc kubenswrapper[4960]: I1201 19:56:23.966541 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" event={"ID":"8b72f468-b777-407c-8e36-30cc398644a6","Type":"ContainerStarted","Data":"4bdec3e3c8be4e09ff98da1d76e832f7ae8fbaf4bad28ae0ff239e4acef7961b"} Dec 01 19:56:23 crc kubenswrapper[4960]: I1201 19:56:23.967597 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:23 crc kubenswrapper[4960]: I1201 19:56:23.967973 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" event={"ID":"914a4b4c-d013-45ef-832d-35b600874887","Type":"ContainerStarted","Data":"42610584bbd7036757d373e1eced3589fd61c1ff18d5e7ef3731fe17486e9823"} Dec 01 19:56:23 crc kubenswrapper[4960]: I1201 19:56:23.968290 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:23 crc kubenswrapper[4960]: I1201 19:56:23.989844 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" podStartSLOduration=2.9954979760000002 podStartE2EDuration="5.989807004s" podCreationTimestamp="2025-12-01 19:56:18 +0000 UTC" firstStartedPulling="2025-12-01 19:56:18.596721315 +0000 UTC m=+1013.884212984" lastFinishedPulling="2025-12-01 19:56:21.591030333 +0000 UTC m=+1016.878522012" observedRunningTime="2025-12-01 19:56:23.988051059 +0000 UTC m=+1019.275542738" watchObservedRunningTime="2025-12-01 19:56:23.989807004 +0000 UTC m=+1019.277298713" Dec 01 19:56:24 crc kubenswrapper[4960]: I1201 19:56:24.011300 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" podStartSLOduration=1.998161576 podStartE2EDuration="6.011281901s" podCreationTimestamp="2025-12-01 19:56:18 +0000 UTC" firstStartedPulling="2025-12-01 19:56:18.968607306 +0000 UTC m=+1014.256098985" lastFinishedPulling="2025-12-01 19:56:22.981727641 +0000 UTC m=+1018.269219310" observedRunningTime="2025-12-01 19:56:24.009794815 +0000 UTC m=+1019.297286484" watchObservedRunningTime="2025-12-01 19:56:24.011281901 +0000 UTC m=+1019.298773570" Dec 01 19:56:38 crc kubenswrapper[4960]: I1201 19:56:38.524765 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d7c4c664-7fttb" Dec 01 19:56:43 crc kubenswrapper[4960]: I1201 19:56:43.870846 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:56:43 crc kubenswrapper[4960]: I1201 19:56:43.871492 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:56:58 crc kubenswrapper[4960]: I1201 19:56:58.358974 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-64f48564cc-pv444" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.223736 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.224684 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.229239 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hqf77" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.230662 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.235820 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2dvm9"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.283441 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.290043 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.290160 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.303297 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.335183 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qztxp"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.336348 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.339474 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.340353 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.340513 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bq8gl" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.340604 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.358412 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-sm9mw"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.359511 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.363314 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.386828 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-sm9mw"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413246 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-metrics\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413320 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zs7p\" (UniqueName: \"kubernetes.io/projected/d98b4520-1781-455c-976c-259b9225fed7-kube-api-access-5zs7p\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d6c8\" (UniqueName: \"kubernetes.io/projected/c7cef304-b739-45e6-b2c8-c729abfed45d-kube-api-access-4d6c8\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413376 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-reloader\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413395 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-conf\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413427 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d98b4520-1781-455c-976c-259b9225fed7-metrics-certs\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413461 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-sockets\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d98b4520-1781-455c-976c-259b9225fed7-frr-startup\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.413512 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7cef304-b739-45e6-b2c8-c729abfed45d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515177 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25nkj\" (UniqueName: \"kubernetes.io/projected/0c199599-aa8d-4e80-93f4-640a8ff141fa-kube-api-access-25nkj\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515256 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zs7p\" (UniqueName: \"kubernetes.io/projected/d98b4520-1781-455c-976c-259b9225fed7-kube-api-access-5zs7p\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515302 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d6c8\" (UniqueName: \"kubernetes.io/projected/c7cef304-b739-45e6-b2c8-c729abfed45d-kube-api-access-4d6c8\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515329 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-reloader\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515347 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-conf\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515370 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-cert\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515410 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d98b4520-1781-455c-976c-259b9225fed7-metrics-certs\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515438 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515478 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-sockets\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515508 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d98b4520-1781-455c-976c-259b9225fed7-frr-startup\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515550 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7cef304-b739-45e6-b2c8-c729abfed45d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515576 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm7mh\" (UniqueName: \"kubernetes.io/projected/41410416-def6-44f6-9a25-14a6b9a50cec-kube-api-access-gm7mh\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515596 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-metrics-certs\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-metrics\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.515638 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0c199599-aa8d-4e80-93f4-640a8ff141fa-metallb-excludel2\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.516687 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-reloader\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.516915 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-conf\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.517891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-frr-sockets\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.518203 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d98b4520-1781-455c-976c-259b9225fed7-metrics\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.518658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d98b4520-1781-455c-976c-259b9225fed7-frr-startup\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.524551 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7cef304-b739-45e6-b2c8-c729abfed45d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.526225 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d98b4520-1781-455c-976c-259b9225fed7-metrics-certs\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.536120 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zs7p\" (UniqueName: \"kubernetes.io/projected/d98b4520-1781-455c-976c-259b9225fed7-kube-api-access-5zs7p\") pod \"frr-k8s-2dvm9\" (UID: \"d98b4520-1781-455c-976c-259b9225fed7\") " pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.537857 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d6c8\" (UniqueName: \"kubernetes.io/projected/c7cef304-b739-45e6-b2c8-c729abfed45d-kube-api-access-4d6c8\") pod \"frr-k8s-webhook-server-7fcb986d4-wjq2w\" (UID: \"c7cef304-b739-45e6-b2c8-c729abfed45d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.551613 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617278 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-cert\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617774 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617822 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617903 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-metrics-certs\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm7mh\" (UniqueName: \"kubernetes.io/projected/41410416-def6-44f6-9a25-14a6b9a50cec-kube-api-access-gm7mh\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617961 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0c199599-aa8d-4e80-93f4-640a8ff141fa-metallb-excludel2\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.617995 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25nkj\" (UniqueName: \"kubernetes.io/projected/0c199599-aa8d-4e80-93f4-640a8ff141fa-kube-api-access-25nkj\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: E1201 19:56:59.618233 4960 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 19:56:59 crc kubenswrapper[4960]: E1201 19:56:59.618312 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist podName:0c199599-aa8d-4e80-93f4-640a8ff141fa nodeName:}" failed. No retries permitted until 2025-12-01 19:57:00.118285102 +0000 UTC m=+1055.405776771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist") pod "speaker-qztxp" (UID: "0c199599-aa8d-4e80-93f4-640a8ff141fa") : secret "metallb-memberlist" not found Dec 01 19:56:59 crc kubenswrapper[4960]: E1201 19:56:59.618571 4960 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 01 19:56:59 crc kubenswrapper[4960]: E1201 19:56:59.618640 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs podName:0c199599-aa8d-4e80-93f4-640a8ff141fa nodeName:}" failed. No retries permitted until 2025-12-01 19:57:00.118611113 +0000 UTC m=+1055.406103002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs") pod "speaker-qztxp" (UID: "0c199599-aa8d-4e80-93f4-640a8ff141fa") : secret "speaker-certs-secret" not found Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.619799 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0c199599-aa8d-4e80-93f4-640a8ff141fa-metallb-excludel2\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.623763 4960 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.625389 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-metrics-certs\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.640387 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41410416-def6-44f6-9a25-14a6b9a50cec-cert\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.641385 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25nkj\" (UniqueName: \"kubernetes.io/projected/0c199599-aa8d-4e80-93f4-640a8ff141fa-kube-api-access-25nkj\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.646005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm7mh\" (UniqueName: \"kubernetes.io/projected/41410416-def6-44f6-9a25-14a6b9a50cec-kube-api-access-gm7mh\") pod \"controller-f8648f98b-sm9mw\" (UID: \"41410416-def6-44f6-9a25-14a6b9a50cec\") " pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.683340 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.797431 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w"] Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.804341 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 19:56:59 crc kubenswrapper[4960]: I1201 19:56:59.896335 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-sm9mw"] Dec 01 19:56:59 crc kubenswrapper[4960]: W1201 19:56:59.898694 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41410416_def6_44f6_9a25_14a6b9a50cec.slice/crio-d63345e541f00066b03adc3c4bf044b6d1fa68c18dd1a1deda4737bd115d6b3f WatchSource:0}: Error finding container d63345e541f00066b03adc3c4bf044b6d1fa68c18dd1a1deda4737bd115d6b3f: Status 404 returned error can't find the container with id d63345e541f00066b03adc3c4bf044b6d1fa68c18dd1a1deda4737bd115d6b3f Dec 01 19:57:00 crc kubenswrapper[4960]: I1201 19:57:00.128514 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:57:00 crc kubenswrapper[4960]: I1201 19:57:00.128597 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:57:00 crc kubenswrapper[4960]: E1201 19:57:00.128762 4960 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 01 19:57:00 crc kubenswrapper[4960]: E1201 19:57:00.128883 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist podName:0c199599-aa8d-4e80-93f4-640a8ff141fa nodeName:}" failed. No retries permitted until 2025-12-01 19:57:01.128857288 +0000 UTC m=+1056.416348957 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist") pod "speaker-qztxp" (UID: "0c199599-aa8d-4e80-93f4-640a8ff141fa") : secret "metallb-memberlist" not found Dec 01 19:57:00 crc kubenswrapper[4960]: I1201 19:57:00.138243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-metrics-certs\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:57:00 crc kubenswrapper[4960]: I1201 19:57:00.296109 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sm9mw" event={"ID":"41410416-def6-44f6-9a25-14a6b9a50cec","Type":"ContainerStarted","Data":"d63345e541f00066b03adc3c4bf044b6d1fa68c18dd1a1deda4737bd115d6b3f"} Dec 01 19:57:00 crc kubenswrapper[4960]: I1201 19:57:00.297093 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" event={"ID":"c7cef304-b739-45e6-b2c8-c729abfed45d","Type":"ContainerStarted","Data":"001f6926145512a218b34486464341450450c6cd3d3c986a974f5727aeb85a0b"} Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.143819 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.177888 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0c199599-aa8d-4e80-93f4-640a8ff141fa-memberlist\") pod \"speaker-qztxp\" (UID: \"0c199599-aa8d-4e80-93f4-640a8ff141fa\") " pod="metallb-system/speaker-qztxp" Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.309049 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"a0acfc5fc2013f272acdd3503bd9a5516924db91b34202bb3edc20a2e7729377"} Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.313606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sm9mw" event={"ID":"41410416-def6-44f6-9a25-14a6b9a50cec","Type":"ContainerStarted","Data":"5c98273c2e7d0fbfeeabdd7e4cf8517bfead11c91d469e5fd3abadf1dd743fd7"} Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.313665 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sm9mw" event={"ID":"41410416-def6-44f6-9a25-14a6b9a50cec","Type":"ContainerStarted","Data":"9e9932da82fda8d579ae214b778a17e8616c2b5337a8257e9177414b144c7912"} Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.314869 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.348438 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-sm9mw" podStartSLOduration=2.348412105 podStartE2EDuration="2.348412105s" podCreationTimestamp="2025-12-01 19:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:57:01.347806576 +0000 UTC m=+1056.635298245" watchObservedRunningTime="2025-12-01 19:57:01.348412105 +0000 UTC m=+1056.635903774" Dec 01 19:57:01 crc kubenswrapper[4960]: I1201 19:57:01.462962 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qztxp" Dec 01 19:57:01 crc kubenswrapper[4960]: W1201 19:57:01.497306 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c199599_aa8d_4e80_93f4_640a8ff141fa.slice/crio-7183aa625c426b503b826df0d0a2bfdd8ac06a72e771dc76e51fa90127b846bc WatchSource:0}: Error finding container 7183aa625c426b503b826df0d0a2bfdd8ac06a72e771dc76e51fa90127b846bc: Status 404 returned error can't find the container with id 7183aa625c426b503b826df0d0a2bfdd8ac06a72e771dc76e51fa90127b846bc Dec 01 19:57:02 crc kubenswrapper[4960]: I1201 19:57:02.329429 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qztxp" event={"ID":"0c199599-aa8d-4e80-93f4-640a8ff141fa","Type":"ContainerStarted","Data":"f233b625e86f9e56741ebb2771c3e3fcbcdebd311d836a330be8b2eafefaa6d5"} Dec 01 19:57:02 crc kubenswrapper[4960]: I1201 19:57:02.329849 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qztxp" event={"ID":"0c199599-aa8d-4e80-93f4-640a8ff141fa","Type":"ContainerStarted","Data":"d628bba29b6063b0e0fde6d617da3865a394d418826e08b9764ee90553aaa29a"} Dec 01 19:57:02 crc kubenswrapper[4960]: I1201 19:57:02.329864 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qztxp" event={"ID":"0c199599-aa8d-4e80-93f4-640a8ff141fa","Type":"ContainerStarted","Data":"7183aa625c426b503b826df0d0a2bfdd8ac06a72e771dc76e51fa90127b846bc"} Dec 01 19:57:02 crc kubenswrapper[4960]: I1201 19:57:02.330072 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qztxp" Dec 01 19:57:02 crc kubenswrapper[4960]: I1201 19:57:02.366907 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qztxp" podStartSLOduration=3.366884131 podStartE2EDuration="3.366884131s" podCreationTimestamp="2025-12-01 19:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:57:02.363902919 +0000 UTC m=+1057.651394588" watchObservedRunningTime="2025-12-01 19:57:02.366884131 +0000 UTC m=+1057.654375800" Dec 01 19:57:08 crc kubenswrapper[4960]: I1201 19:57:08.398963 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" event={"ID":"c7cef304-b739-45e6-b2c8-c729abfed45d","Type":"ContainerStarted","Data":"9192812f8ddfdf0f6ee0879ca1c21f4250e181fc08bed99dc86454e9980ec3e2"} Dec 01 19:57:08 crc kubenswrapper[4960]: I1201 19:57:08.399724 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:57:08 crc kubenswrapper[4960]: I1201 19:57:08.401146 4960 generic.go:334] "Generic (PLEG): container finished" podID="d98b4520-1781-455c-976c-259b9225fed7" containerID="a19360c5b78917745ee9934b41003ef1c4a3e96c273c4e99f5af7d523598df9b" exitCode=0 Dec 01 19:57:08 crc kubenswrapper[4960]: I1201 19:57:08.401206 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerDied","Data":"a19360c5b78917745ee9934b41003ef1c4a3e96c273c4e99f5af7d523598df9b"} Dec 01 19:57:08 crc kubenswrapper[4960]: I1201 19:57:08.426850 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" podStartSLOduration=2.110434841 podStartE2EDuration="9.426822325s" podCreationTimestamp="2025-12-01 19:56:59 +0000 UTC" firstStartedPulling="2025-12-01 19:56:59.804053698 +0000 UTC m=+1055.091545367" lastFinishedPulling="2025-12-01 19:57:07.120441162 +0000 UTC m=+1062.407932851" observedRunningTime="2025-12-01 19:57:08.425061069 +0000 UTC m=+1063.712552758" watchObservedRunningTime="2025-12-01 19:57:08.426822325 +0000 UTC m=+1063.714313994" Dec 01 19:57:09 crc kubenswrapper[4960]: I1201 19:57:09.413263 4960 generic.go:334] "Generic (PLEG): container finished" podID="d98b4520-1781-455c-976c-259b9225fed7" containerID="e3560663a3a93f15441b61d64a87ccf429d484a11a34ae8d580574ee6ec054f3" exitCode=0 Dec 01 19:57:09 crc kubenswrapper[4960]: I1201 19:57:09.413323 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerDied","Data":"e3560663a3a93f15441b61d64a87ccf429d484a11a34ae8d580574ee6ec054f3"} Dec 01 19:57:10 crc kubenswrapper[4960]: I1201 19:57:10.424956 4960 generic.go:334] "Generic (PLEG): container finished" podID="d98b4520-1781-455c-976c-259b9225fed7" containerID="a869dc6b06e79c08112314f945fdee96fe52cecb7f20e49af9f7bca18d19637a" exitCode=0 Dec 01 19:57:10 crc kubenswrapper[4960]: I1201 19:57:10.425167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerDied","Data":"a869dc6b06e79c08112314f945fdee96fe52cecb7f20e49af9f7bca18d19637a"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.444956 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"5220ce3aa211b68a7727514c24282fe10891d32c9b8cbcb9474f6151a4a79f25"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.445440 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"059731ae46931e3dab9288da172912c020aeb7f38f462d720dc0d41b08f9b17e"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.445456 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"a06e9bb6c2da6c15100b425be27efad9cfd4c95283372d95d61c58bc2e19126e"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.445470 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"d4e28860eac4b94ebfb64d80cf0621e2a2eeae37ce8fe54aef5da06166461425"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.445482 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"c9d98c8e0df1c286b9e942929547780e2aca9beda70bf4cbabb3b9bd971be448"} Dec 01 19:57:11 crc kubenswrapper[4960]: I1201 19:57:11.467395 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qztxp" Dec 01 19:57:12 crc kubenswrapper[4960]: I1201 19:57:12.459869 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2dvm9" event={"ID":"d98b4520-1781-455c-976c-259b9225fed7","Type":"ContainerStarted","Data":"a048110f6676c887e10086f8af40a8f6bf3770cbdb00ffaf8763983c5dad1747"} Dec 01 19:57:12 crc kubenswrapper[4960]: I1201 19:57:12.460540 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:57:12 crc kubenswrapper[4960]: I1201 19:57:12.498689 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2dvm9" podStartSLOduration=6.857201289 podStartE2EDuration="13.498653638s" podCreationTimestamp="2025-12-01 19:56:59 +0000 UTC" firstStartedPulling="2025-12-01 19:57:00.507725245 +0000 UTC m=+1055.795216924" lastFinishedPulling="2025-12-01 19:57:07.149177604 +0000 UTC m=+1062.436669273" observedRunningTime="2025-12-01 19:57:12.492801497 +0000 UTC m=+1067.780293206" watchObservedRunningTime="2025-12-01 19:57:12.498653638 +0000 UTC m=+1067.786145347" Dec 01 19:57:13 crc kubenswrapper[4960]: I1201 19:57:13.870605 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:57:13 crc kubenswrapper[4960]: I1201 19:57:13.871202 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:57:13 crc kubenswrapper[4960]: I1201 19:57:13.871256 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 19:57:13 crc kubenswrapper[4960]: I1201 19:57:13.871816 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 19:57:13 crc kubenswrapper[4960]: I1201 19:57:13.871880 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09" gracePeriod=600 Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.487502 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09" exitCode=0 Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.487754 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09"} Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.487952 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07"} Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.487980 4960 scope.go:117] "RemoveContainer" containerID="6a46f4296788e41ead6ec45c10b68b5aeacb027d305ea2e3f749c89a78fc801d" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.619233 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.703734 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.964525 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.965809 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.967928 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.968365 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.968395 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5vwhz" Dec 01 19:57:14 crc kubenswrapper[4960]: I1201 19:57:14.994147 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:15 crc kubenswrapper[4960]: I1201 19:57:15.083841 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68lbn\" (UniqueName: \"kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn\") pod \"openstack-operator-index-n69ff\" (UID: \"a18e4581-1629-4efe-bf92-586294bdab57\") " pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:15 crc kubenswrapper[4960]: I1201 19:57:15.185091 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68lbn\" (UniqueName: \"kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn\") pod \"openstack-operator-index-n69ff\" (UID: \"a18e4581-1629-4efe-bf92-586294bdab57\") " pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:15 crc kubenswrapper[4960]: I1201 19:57:15.206568 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68lbn\" (UniqueName: \"kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn\") pod \"openstack-operator-index-n69ff\" (UID: \"a18e4581-1629-4efe-bf92-586294bdab57\") " pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:15 crc kubenswrapper[4960]: I1201 19:57:15.292165 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:15 crc kubenswrapper[4960]: I1201 19:57:15.710572 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:15 crc kubenswrapper[4960]: W1201 19:57:15.723282 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda18e4581_1629_4efe_bf92_586294bdab57.slice/crio-5e03538a6997464b8a625d0e25b63f157f0c6638125e433913499cf2a5492c89 WatchSource:0}: Error finding container 5e03538a6997464b8a625d0e25b63f157f0c6638125e433913499cf2a5492c89: Status 404 returned error can't find the container with id 5e03538a6997464b8a625d0e25b63f157f0c6638125e433913499cf2a5492c89 Dec 01 19:57:16 crc kubenswrapper[4960]: I1201 19:57:16.506368 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n69ff" event={"ID":"a18e4581-1629-4efe-bf92-586294bdab57","Type":"ContainerStarted","Data":"5e03538a6997464b8a625d0e25b63f157f0c6638125e433913499cf2a5492c89"} Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.349028 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.525839 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n69ff" event={"ID":"a18e4581-1629-4efe-bf92-586294bdab57","Type":"ContainerStarted","Data":"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98"} Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.551047 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-n69ff" podStartSLOduration=2.263403703 podStartE2EDuration="4.551011706s" podCreationTimestamp="2025-12-01 19:57:14 +0000 UTC" firstStartedPulling="2025-12-01 19:57:15.727372998 +0000 UTC m=+1071.014864707" lastFinishedPulling="2025-12-01 19:57:18.014981041 +0000 UTC m=+1073.302472710" observedRunningTime="2025-12-01 19:57:18.549161799 +0000 UTC m=+1073.836653498" watchObservedRunningTime="2025-12-01 19:57:18.551011706 +0000 UTC m=+1073.838503415" Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.958445 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2jktp"] Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.960977 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:18 crc kubenswrapper[4960]: I1201 19:57:18.966515 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jktp"] Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.146022 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tjwq\" (UniqueName: \"kubernetes.io/projected/d683e2dd-4bba-4833-ae18-4636090ef1e5-kube-api-access-5tjwq\") pod \"openstack-operator-index-2jktp\" (UID: \"d683e2dd-4bba-4833-ae18-4636090ef1e5\") " pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.247658 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tjwq\" (UniqueName: \"kubernetes.io/projected/d683e2dd-4bba-4833-ae18-4636090ef1e5-kube-api-access-5tjwq\") pod \"openstack-operator-index-2jktp\" (UID: \"d683e2dd-4bba-4833-ae18-4636090ef1e5\") " pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.272390 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tjwq\" (UniqueName: \"kubernetes.io/projected/d683e2dd-4bba-4833-ae18-4636090ef1e5-kube-api-access-5tjwq\") pod \"openstack-operator-index-2jktp\" (UID: \"d683e2dd-4bba-4833-ae18-4636090ef1e5\") " pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.286744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.541353 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-n69ff" podUID="a18e4581-1629-4efe-bf92-586294bdab57" containerName="registry-server" containerID="cri-o://d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98" gracePeriod=2 Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.568372 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wjq2w" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.688975 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-sm9mw" Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.790654 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jktp"] Dec 01 19:57:19 crc kubenswrapper[4960]: W1201 19:57:19.811307 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd683e2dd_4bba_4833_ae18_4636090ef1e5.slice/crio-19d827226fedcdbf92ecb4d847d38cd417103762d9dc069c690eaa8d49d371f0 WatchSource:0}: Error finding container 19d827226fedcdbf92ecb4d847d38cd417103762d9dc069c690eaa8d49d371f0: Status 404 returned error can't find the container with id 19d827226fedcdbf92ecb4d847d38cd417103762d9dc069c690eaa8d49d371f0 Dec 01 19:57:19 crc kubenswrapper[4960]: I1201 19:57:19.916645 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.061377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68lbn\" (UniqueName: \"kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn\") pod \"a18e4581-1629-4efe-bf92-586294bdab57\" (UID: \"a18e4581-1629-4efe-bf92-586294bdab57\") " Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.067739 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn" (OuterVolumeSpecName: "kube-api-access-68lbn") pod "a18e4581-1629-4efe-bf92-586294bdab57" (UID: "a18e4581-1629-4efe-bf92-586294bdab57"). InnerVolumeSpecName "kube-api-access-68lbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.163345 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68lbn\" (UniqueName: \"kubernetes.io/projected/a18e4581-1629-4efe-bf92-586294bdab57-kube-api-access-68lbn\") on node \"crc\" DevicePath \"\"" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.551529 4960 generic.go:334] "Generic (PLEG): container finished" podID="a18e4581-1629-4efe-bf92-586294bdab57" containerID="d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98" exitCode=0 Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.551623 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-n69ff" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.551625 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n69ff" event={"ID":"a18e4581-1629-4efe-bf92-586294bdab57","Type":"ContainerDied","Data":"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98"} Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.551729 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-n69ff" event={"ID":"a18e4581-1629-4efe-bf92-586294bdab57","Type":"ContainerDied","Data":"5e03538a6997464b8a625d0e25b63f157f0c6638125e433913499cf2a5492c89"} Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.551767 4960 scope.go:117] "RemoveContainer" containerID="d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.555496 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jktp" event={"ID":"d683e2dd-4bba-4833-ae18-4636090ef1e5","Type":"ContainerStarted","Data":"0dfc6cd72a188c2dd51eaf5ead7e260a9cd780d1310ea809cba422283220ae7d"} Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.555562 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jktp" event={"ID":"d683e2dd-4bba-4833-ae18-4636090ef1e5","Type":"ContainerStarted","Data":"19d827226fedcdbf92ecb4d847d38cd417103762d9dc069c690eaa8d49d371f0"} Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.585268 4960 scope.go:117] "RemoveContainer" containerID="d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.587292 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2jktp" podStartSLOduration=2.527905647 podStartE2EDuration="2.587257399s" podCreationTimestamp="2025-12-01 19:57:18 +0000 UTC" firstStartedPulling="2025-12-01 19:57:19.830000988 +0000 UTC m=+1075.117492667" lastFinishedPulling="2025-12-01 19:57:19.88935276 +0000 UTC m=+1075.176844419" observedRunningTime="2025-12-01 19:57:20.581579203 +0000 UTC m=+1075.869070892" watchObservedRunningTime="2025-12-01 19:57:20.587257399 +0000 UTC m=+1075.874749108" Dec 01 19:57:20 crc kubenswrapper[4960]: E1201 19:57:20.587592 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98\": container with ID starting with d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98 not found: ID does not exist" containerID="d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.587646 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98"} err="failed to get container status \"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98\": rpc error: code = NotFound desc = could not find container \"d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98\": container with ID starting with d3f22de55fa9e4d1f8254abfaa0ab61985c8df29f7212cc4c8050a76e6215f98 not found: ID does not exist" Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.609497 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:20 crc kubenswrapper[4960]: I1201 19:57:20.619760 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-n69ff"] Dec 01 19:57:21 crc kubenswrapper[4960]: I1201 19:57:21.341202 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a18e4581-1629-4efe-bf92-586294bdab57" path="/var/lib/kubelet/pods/a18e4581-1629-4efe-bf92-586294bdab57/volumes" Dec 01 19:57:29 crc kubenswrapper[4960]: I1201 19:57:29.287760 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:29 crc kubenswrapper[4960]: I1201 19:57:29.288357 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:29 crc kubenswrapper[4960]: I1201 19:57:29.334956 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:29 crc kubenswrapper[4960]: I1201 19:57:29.622848 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2dvm9" Dec 01 19:57:29 crc kubenswrapper[4960]: I1201 19:57:29.678182 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2jktp" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.395847 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz"] Dec 01 19:57:31 crc kubenswrapper[4960]: E1201 19:57:31.396584 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a18e4581-1629-4efe-bf92-586294bdab57" containerName="registry-server" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.396598 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="a18e4581-1629-4efe-bf92-586294bdab57" containerName="registry-server" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.396726 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="a18e4581-1629-4efe-bf92-586294bdab57" containerName="registry-server" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.397594 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.400682 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-cf8h7" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.409038 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz"] Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.539892 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.540006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.540072 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk82m\" (UniqueName: \"kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.641851 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.641988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.642039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk82m\" (UniqueName: \"kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.642737 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.642888 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.677997 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk82m\" (UniqueName: \"kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m\") pod \"2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:31 crc kubenswrapper[4960]: I1201 19:57:31.718994 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:32 crc kubenswrapper[4960]: I1201 19:57:32.184313 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz"] Dec 01 19:57:32 crc kubenswrapper[4960]: W1201 19:57:32.195406 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87988f3c_1a59_4ebd_b80f_c590bd7e4e57.slice/crio-4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9 WatchSource:0}: Error finding container 4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9: Status 404 returned error can't find the container with id 4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9 Dec 01 19:57:32 crc kubenswrapper[4960]: I1201 19:57:32.662557 4960 generic.go:334] "Generic (PLEG): container finished" podID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerID="e38018836ceea57cd3ace9567c3519582921e6f5aeace927ec57add788dc14e2" exitCode=0 Dec 01 19:57:32 crc kubenswrapper[4960]: I1201 19:57:32.662625 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" event={"ID":"87988f3c-1a59-4ebd-b80f-c590bd7e4e57","Type":"ContainerDied","Data":"e38018836ceea57cd3ace9567c3519582921e6f5aeace927ec57add788dc14e2"} Dec 01 19:57:32 crc kubenswrapper[4960]: I1201 19:57:32.663156 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" event={"ID":"87988f3c-1a59-4ebd-b80f-c590bd7e4e57","Type":"ContainerStarted","Data":"4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9"} Dec 01 19:57:33 crc kubenswrapper[4960]: I1201 19:57:33.675489 4960 generic.go:334] "Generic (PLEG): container finished" podID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerID="3ae0ae879727df83e72a4b1bf5939255565974a8579fbaec75e8123e6d5fea20" exitCode=0 Dec 01 19:57:33 crc kubenswrapper[4960]: I1201 19:57:33.675572 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" event={"ID":"87988f3c-1a59-4ebd-b80f-c590bd7e4e57","Type":"ContainerDied","Data":"3ae0ae879727df83e72a4b1bf5939255565974a8579fbaec75e8123e6d5fea20"} Dec 01 19:57:34 crc kubenswrapper[4960]: I1201 19:57:34.685256 4960 generic.go:334] "Generic (PLEG): container finished" podID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerID="a68bc5583724cb3b6929702c06dd4faf47cf65c18db68da46e44d63039cc8076" exitCode=0 Dec 01 19:57:34 crc kubenswrapper[4960]: I1201 19:57:34.685568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" event={"ID":"87988f3c-1a59-4ebd-b80f-c590bd7e4e57","Type":"ContainerDied","Data":"a68bc5583724cb3b6929702c06dd4faf47cf65c18db68da46e44d63039cc8076"} Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.031087 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.207939 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle\") pod \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.208095 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk82m\" (UniqueName: \"kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m\") pod \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.208293 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util\") pod \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\" (UID: \"87988f3c-1a59-4ebd-b80f-c590bd7e4e57\") " Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.209297 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle" (OuterVolumeSpecName: "bundle") pod "87988f3c-1a59-4ebd-b80f-c590bd7e4e57" (UID: "87988f3c-1a59-4ebd-b80f-c590bd7e4e57"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.217583 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m" (OuterVolumeSpecName: "kube-api-access-zk82m") pod "87988f3c-1a59-4ebd-b80f-c590bd7e4e57" (UID: "87988f3c-1a59-4ebd-b80f-c590bd7e4e57"). InnerVolumeSpecName "kube-api-access-zk82m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.221626 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util" (OuterVolumeSpecName: "util") pod "87988f3c-1a59-4ebd-b80f-c590bd7e4e57" (UID: "87988f3c-1a59-4ebd-b80f-c590bd7e4e57"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.309826 4960 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-util\") on node \"crc\" DevicePath \"\"" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.309897 4960 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.309913 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk82m\" (UniqueName: \"kubernetes.io/projected/87988f3c-1a59-4ebd-b80f-c590bd7e4e57-kube-api-access-zk82m\") on node \"crc\" DevicePath \"\"" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.704468 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" event={"ID":"87988f3c-1a59-4ebd-b80f-c590bd7e4e57","Type":"ContainerDied","Data":"4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9"} Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.704849 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c66744c0db56cc0d756d79a7180e49da6e3aadba6a80223fa0d81ddb78481b9" Dec 01 19:57:36 crc kubenswrapper[4960]: I1201 19:57:36.704923 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.309335 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8"] Dec 01 19:57:43 crc kubenswrapper[4960]: E1201 19:57:43.310327 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="util" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.310343 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="util" Dec 01 19:57:43 crc kubenswrapper[4960]: E1201 19:57:43.310388 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="extract" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.310395 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="extract" Dec 01 19:57:43 crc kubenswrapper[4960]: E1201 19:57:43.310409 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="pull" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.310416 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="pull" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.310556 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="87988f3c-1a59-4ebd-b80f-c590bd7e4e57" containerName="extract" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.311156 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.314139 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-p6755" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.335172 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8"] Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.421141 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5txj7\" (UniqueName: \"kubernetes.io/projected/9179ddbd-9397-405d-93d5-e32ce4963305-kube-api-access-5txj7\") pod \"openstack-operator-controller-operator-8455849bfc-vgkq8\" (UID: \"9179ddbd-9397-405d-93d5-e32ce4963305\") " pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.523183 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5txj7\" (UniqueName: \"kubernetes.io/projected/9179ddbd-9397-405d-93d5-e32ce4963305-kube-api-access-5txj7\") pod \"openstack-operator-controller-operator-8455849bfc-vgkq8\" (UID: \"9179ddbd-9397-405d-93d5-e32ce4963305\") " pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.549445 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5txj7\" (UniqueName: \"kubernetes.io/projected/9179ddbd-9397-405d-93d5-e32ce4963305-kube-api-access-5txj7\") pod \"openstack-operator-controller-operator-8455849bfc-vgkq8\" (UID: \"9179ddbd-9397-405d-93d5-e32ce4963305\") " pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:43 crc kubenswrapper[4960]: I1201 19:57:43.629617 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:44 crc kubenswrapper[4960]: I1201 19:57:44.063687 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8"] Dec 01 19:57:44 crc kubenswrapper[4960]: I1201 19:57:44.771680 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" event={"ID":"9179ddbd-9397-405d-93d5-e32ce4963305","Type":"ContainerStarted","Data":"dd566a7515fc9a72a1b27ffa3180e320a8651169b9a24db5d28d9f3db13960b2"} Dec 01 19:57:48 crc kubenswrapper[4960]: I1201 19:57:48.804023 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" event={"ID":"9179ddbd-9397-405d-93d5-e32ce4963305","Type":"ContainerStarted","Data":"bcc78471b07c5ad533622bfdceff3da3f969f405d3731566fb24e2cc8e30c637"} Dec 01 19:57:48 crc kubenswrapper[4960]: I1201 19:57:48.804957 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:57:48 crc kubenswrapper[4960]: I1201 19:57:48.839404 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" podStartSLOduration=1.343768195 podStartE2EDuration="5.839384444s" podCreationTimestamp="2025-12-01 19:57:43 +0000 UTC" firstStartedPulling="2025-12-01 19:57:44.075988123 +0000 UTC m=+1099.363479802" lastFinishedPulling="2025-12-01 19:57:48.571604372 +0000 UTC m=+1103.859096051" observedRunningTime="2025-12-01 19:57:48.834233723 +0000 UTC m=+1104.121725392" watchObservedRunningTime="2025-12-01 19:57:48.839384444 +0000 UTC m=+1104.126876113" Dec 01 19:57:53 crc kubenswrapper[4960]: I1201 19:57:53.633048 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8455849bfc-vgkq8" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.606527 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.610967 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.614897 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wnhsm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.619617 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.621060 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.624793 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.630370 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ttxzv" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.637107 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.650168 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.651414 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.659602 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-22s7s" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.670228 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.671602 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.676579 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-qs4j9" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.692831 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.707215 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.719103 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.721684 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.724614 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-nk2hm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.728655 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.730239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.737635 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lsl8l" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.751182 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.764146 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzl7x\" (UniqueName: \"kubernetes.io/projected/238b3987-76a6-401d-a33c-5c2efd1bafcf-kube-api-access-xzl7x\") pod \"barbican-operator-controller-manager-7d9dfd778-8l7sf\" (UID: \"238b3987-76a6-401d-a33c-5c2efd1bafcf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.764228 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvl9x\" (UniqueName: \"kubernetes.io/projected/a51010d5-7158-47ce-9428-25f5b2ca871f-kube-api-access-wvl9x\") pod \"designate-operator-controller-manager-78b4bc895b-6tbkt\" (UID: \"a51010d5-7158-47ce-9428-25f5b2ca871f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.764264 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478zm\" (UniqueName: \"kubernetes.io/projected/018cf742-9a47-4355-879a-9e064f358748-kube-api-access-478zm\") pod \"glance-operator-controller-manager-668d9c48b9-pbxhm\" (UID: \"018cf742-9a47-4355-879a-9e064f358748\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.764286 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldv5b\" (UniqueName: \"kubernetes.io/projected/cce3a29f-5414-4fee-a27c-9b6286930903-kube-api-access-ldv5b\") pod \"cinder-operator-controller-manager-859b6ccc6-p5qfh\" (UID: \"cce3a29f-5414-4fee-a27c-9b6286930903\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.764370 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.787089 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.811451 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.864145 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hlx5s" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.865929 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478zm\" (UniqueName: \"kubernetes.io/projected/018cf742-9a47-4355-879a-9e064f358748-kube-api-access-478zm\") pod \"glance-operator-controller-manager-668d9c48b9-pbxhm\" (UID: \"018cf742-9a47-4355-879a-9e064f358748\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.866002 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldv5b\" (UniqueName: \"kubernetes.io/projected/cce3a29f-5414-4fee-a27c-9b6286930903-kube-api-access-ldv5b\") pod \"cinder-operator-controller-manager-859b6ccc6-p5qfh\" (UID: \"cce3a29f-5414-4fee-a27c-9b6286930903\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.866039 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp72m\" (UniqueName: \"kubernetes.io/projected/0a07e8af-0ea2-4b3e-b9b7-b01f88025904-kube-api-access-sp72m\") pod \"heat-operator-controller-manager-5f64f6f8bb-6mqj7\" (UID: \"0a07e8af-0ea2-4b3e-b9b7-b01f88025904\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.866074 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpltw\" (UniqueName: \"kubernetes.io/projected/dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d-kube-api-access-lpltw\") pod \"horizon-operator-controller-manager-68c6d99b8f-bsfbn\" (UID: \"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.866107 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzl7x\" (UniqueName: \"kubernetes.io/projected/238b3987-76a6-401d-a33c-5c2efd1bafcf-kube-api-access-xzl7x\") pod \"barbican-operator-controller-manager-7d9dfd778-8l7sf\" (UID: \"238b3987-76a6-401d-a33c-5c2efd1bafcf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.866172 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvl9x\" (UniqueName: \"kubernetes.io/projected/a51010d5-7158-47ce-9428-25f5b2ca871f-kube-api-access-wvl9x\") pod \"designate-operator-controller-manager-78b4bc895b-6tbkt\" (UID: \"a51010d5-7158-47ce-9428-25f5b2ca871f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.878001 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.878995 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.879504 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.885557 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-zv7sr" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.888914 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.900726 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.902414 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.910943 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.912373 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-t44v2" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.914581 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzl7x\" (UniqueName: \"kubernetes.io/projected/238b3987-76a6-401d-a33c-5c2efd1bafcf-kube-api-access-xzl7x\") pod \"barbican-operator-controller-manager-7d9dfd778-8l7sf\" (UID: \"238b3987-76a6-401d-a33c-5c2efd1bafcf\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.928015 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldv5b\" (UniqueName: \"kubernetes.io/projected/cce3a29f-5414-4fee-a27c-9b6286930903-kube-api-access-ldv5b\") pod \"cinder-operator-controller-manager-859b6ccc6-p5qfh\" (UID: \"cce3a29f-5414-4fee-a27c-9b6286930903\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.928787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvl9x\" (UniqueName: \"kubernetes.io/projected/a51010d5-7158-47ce-9428-25f5b2ca871f-kube-api-access-wvl9x\") pod \"designate-operator-controller-manager-78b4bc895b-6tbkt\" (UID: \"a51010d5-7158-47ce-9428-25f5b2ca871f\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.931756 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.932787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478zm\" (UniqueName: \"kubernetes.io/projected/018cf742-9a47-4355-879a-9e064f358748-kube-api-access-478zm\") pod \"glance-operator-controller-manager-668d9c48b9-pbxhm\" (UID: \"018cf742-9a47-4355-879a-9e064f358748\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.936190 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.948483 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.951210 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.952841 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.957259 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.959925 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.960619 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-pq84g" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.965907 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9qsp5" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.967016 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.967055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj5z\" (UniqueName: \"kubernetes.io/projected/58815c45-1d27-4f86-9b6a-ffbb04c4e250-kube-api-access-qsj5z\") pod \"ironic-operator-controller-manager-6c548fd776-zggh7\" (UID: \"58815c45-1d27-4f86-9b6a-ffbb04c4e250\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.967084 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp72m\" (UniqueName: \"kubernetes.io/projected/0a07e8af-0ea2-4b3e-b9b7-b01f88025904-kube-api-access-sp72m\") pod \"heat-operator-controller-manager-5f64f6f8bb-6mqj7\" (UID: \"0a07e8af-0ea2-4b3e-b9b7-b01f88025904\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.967133 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf4w4\" (UniqueName: \"kubernetes.io/projected/b801b659-3926-4a9f-a36f-f752a6bd9563-kube-api-access-lf4w4\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.967162 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpltw\" (UniqueName: \"kubernetes.io/projected/dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d-kube-api-access-lpltw\") pod \"horizon-operator-controller-manager-68c6d99b8f-bsfbn\" (UID: \"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.971255 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.971545 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.978329 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.988604 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7"] Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.989738 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.992974 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-lnxvw" Dec 01 19:58:12 crc kubenswrapper[4960]: I1201 19:58:12.993288 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.004044 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.005453 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.010749 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-flgdt" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.026029 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.028718 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpltw\" (UniqueName: \"kubernetes.io/projected/dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d-kube-api-access-lpltw\") pod \"horizon-operator-controller-manager-68c6d99b8f-bsfbn\" (UID: \"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.032890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp72m\" (UniqueName: \"kubernetes.io/projected/0a07e8af-0ea2-4b3e-b9b7-b01f88025904-kube-api-access-sp72m\") pod \"heat-operator-controller-manager-5f64f6f8bb-6mqj7\" (UID: \"0a07e8af-0ea2-4b3e-b9b7-b01f88025904\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.035377 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.045593 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.047387 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.050684 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-7tkbc" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.053408 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.058238 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.060763 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069196 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj5z\" (UniqueName: \"kubernetes.io/projected/58815c45-1d27-4f86-9b6a-ffbb04c4e250-kube-api-access-qsj5z\") pod \"ironic-operator-controller-manager-6c548fd776-zggh7\" (UID: \"58815c45-1d27-4f86-9b6a-ffbb04c4e250\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069227 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb68q\" (UniqueName: \"kubernetes.io/projected/de2501b9-bdd7-46c3-9128-20b2d9d5ab88-kube-api-access-vb68q\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-m24v2\" (UID: \"de2501b9-bdd7-46c3-9128-20b2d9d5ab88\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069250 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf4w4\" (UniqueName: \"kubernetes.io/projected/b801b659-3926-4a9f-a36f-f752a6bd9563-kube-api-access-lf4w4\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069268 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfrhs\" (UniqueName: \"kubernetes.io/projected/8b922299-0b5e-4595-bdab-1c759fd99465-kube-api-access-hfrhs\") pod \"keystone-operator-controller-manager-546d4bdf48-sfsvg\" (UID: \"8b922299-0b5e-4595-bdab-1c759fd99465\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.069302 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5jvp\" (UniqueName: \"kubernetes.io/projected/3b481993-f9fd-4d99-98e3-04b51c186e20-kube-api-access-z5jvp\") pod \"manila-operator-controller-manager-6546668bfd-hcpbb\" (UID: \"3b481993-f9fd-4d99-98e3-04b51c186e20\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.069463 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.069510 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:13.569492258 +0000 UTC m=+1128.856983927 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.077485 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.079073 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.085614 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5gsf7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.085653 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.089325 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.093624 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf4w4\" (UniqueName: \"kubernetes.io/projected/b801b659-3926-4a9f-a36f-f752a6bd9563-kube-api-access-lf4w4\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.094174 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj5z\" (UniqueName: \"kubernetes.io/projected/58815c45-1d27-4f86-9b6a-ffbb04c4e250-kube-api-access-qsj5z\") pod \"ironic-operator-controller-manager-6c548fd776-zggh7\" (UID: \"58815c45-1d27-4f86-9b6a-ffbb04c4e250\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.109455 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.110645 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.113967 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-h4p7h" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.115169 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.125256 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.126918 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.130639 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7mwn4" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.138449 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.145943 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.147790 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.153370 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-844jh" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.154510 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.170733 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5jvp\" (UniqueName: \"kubernetes.io/projected/3b481993-f9fd-4d99-98e3-04b51c186e20-kube-api-access-z5jvp\") pod \"manila-operator-controller-manager-6546668bfd-hcpbb\" (UID: \"3b481993-f9fd-4d99-98e3-04b51c186e20\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.170963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4khj\" (UniqueName: \"kubernetes.io/projected/b869ba57-9812-4684-8cbb-48fc606c6aec-kube-api-access-s4khj\") pod \"octavia-operator-controller-manager-998648c74-cjgcs\" (UID: \"b869ba57-9812-4684-8cbb-48fc606c6aec\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171082 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-559bk\" (UniqueName: \"kubernetes.io/projected/0106714e-857b-4ade-9ee2-66fe8c12164c-kube-api-access-559bk\") pod \"nova-operator-controller-manager-697bc559fc-ld68p\" (UID: \"0106714e-857b-4ade-9ee2-66fe8c12164c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171185 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb68q\" (UniqueName: \"kubernetes.io/projected/de2501b9-bdd7-46c3-9128-20b2d9d5ab88-kube-api-access-vb68q\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-m24v2\" (UID: \"de2501b9-bdd7-46c3-9128-20b2d9d5ab88\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171207 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfrhs\" (UniqueName: \"kubernetes.io/projected/8b922299-0b5e-4595-bdab-1c759fd99465-kube-api-access-hfrhs\") pod \"keystone-operator-controller-manager-546d4bdf48-sfsvg\" (UID: \"8b922299-0b5e-4595-bdab-1c759fd99465\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171229 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47njt\" (UniqueName: \"kubernetes.io/projected/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-kube-api-access-47njt\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.171258 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7glvk\" (UniqueName: \"kubernetes.io/projected/8c0036c6-9f32-4163-866d-c1f5786c5a2e-kube-api-access-7glvk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-q8bv7\" (UID: \"8c0036c6-9f32-4163-866d-c1f5786c5a2e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.173557 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.174813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.195078 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-5wszt" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.216323 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfrhs\" (UniqueName: \"kubernetes.io/projected/8b922299-0b5e-4595-bdab-1c759fd99465-kube-api-access-hfrhs\") pod \"keystone-operator-controller-manager-546d4bdf48-sfsvg\" (UID: \"8b922299-0b5e-4595-bdab-1c759fd99465\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.234866 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb68q\" (UniqueName: \"kubernetes.io/projected/de2501b9-bdd7-46c3-9128-20b2d9d5ab88-kube-api-access-vb68q\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-m24v2\" (UID: \"de2501b9-bdd7-46c3-9128-20b2d9d5ab88\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.270256 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273338 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4khj\" (UniqueName: \"kubernetes.io/projected/b869ba57-9812-4684-8cbb-48fc606c6aec-kube-api-access-s4khj\") pod \"octavia-operator-controller-manager-998648c74-cjgcs\" (UID: \"b869ba57-9812-4684-8cbb-48fc606c6aec\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273415 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-559bk\" (UniqueName: \"kubernetes.io/projected/0106714e-857b-4ade-9ee2-66fe8c12164c-kube-api-access-559bk\") pod \"nova-operator-controller-manager-697bc559fc-ld68p\" (UID: \"0106714e-857b-4ade-9ee2-66fe8c12164c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273442 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnvq6\" (UniqueName: \"kubernetes.io/projected/406103b8-0898-4d8f-badd-5f32648ad242-kube-api-access-vnvq6\") pod \"swift-operator-controller-manager-5f8c65bbfc-6zgq8\" (UID: \"406103b8-0898-4d8f-badd-5f32648ad242\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273518 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t57p\" (UniqueName: \"kubernetes.io/projected/0b693fe0-7de6-4a0c-a931-21174d92c184-kube-api-access-6t57p\") pod \"telemetry-operator-controller-manager-75c4b445dc-tfgqn\" (UID: \"0b693fe0-7de6-4a0c-a931-21174d92c184\") " pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6sds\" (UniqueName: \"kubernetes.io/projected/b6018c51-4633-48dd-97ea-0d3d405f4e14-kube-api-access-n6sds\") pod \"placement-operator-controller-manager-78f8948974-vjb5b\" (UID: \"b6018c51-4633-48dd-97ea-0d3d405f4e14\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273620 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgpsp\" (UniqueName: \"kubernetes.io/projected/9b79fea5-fb83-4804-987d-b61e11999829-kube-api-access-vgpsp\") pod \"ovn-operator-controller-manager-b6456fdb6-m9qpq\" (UID: \"9b79fea5-fb83-4804-987d-b61e11999829\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47njt\" (UniqueName: \"kubernetes.io/projected/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-kube-api-access-47njt\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.273680 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7glvk\" (UniqueName: \"kubernetes.io/projected/8c0036c6-9f32-4163-866d-c1f5786c5a2e-kube-api-access-7glvk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-q8bv7\" (UID: \"8c0036c6-9f32-4163-866d-c1f5786c5a2e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.273781 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.273893 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert podName:6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c nodeName:}" failed. No retries permitted until 2025-12-01 19:58:13.773858404 +0000 UTC m=+1129.061350073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" (UID: "6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.287024 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.320766 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.321010 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5jvp\" (UniqueName: \"kubernetes.io/projected/3b481993-f9fd-4d99-98e3-04b51c186e20-kube-api-access-z5jvp\") pod \"manila-operator-controller-manager-6546668bfd-hcpbb\" (UID: \"3b481993-f9fd-4d99-98e3-04b51c186e20\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.328300 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-559bk\" (UniqueName: \"kubernetes.io/projected/0106714e-857b-4ade-9ee2-66fe8c12164c-kube-api-access-559bk\") pod \"nova-operator-controller-manager-697bc559fc-ld68p\" (UID: \"0106714e-857b-4ade-9ee2-66fe8c12164c\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.328367 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7glvk\" (UniqueName: \"kubernetes.io/projected/8c0036c6-9f32-4163-866d-c1f5786c5a2e-kube-api-access-7glvk\") pod \"mariadb-operator-controller-manager-56bbcc9d85-q8bv7\" (UID: \"8c0036c6-9f32-4163-866d-c1f5786c5a2e\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.330644 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4khj\" (UniqueName: \"kubernetes.io/projected/b869ba57-9812-4684-8cbb-48fc606c6aec-kube-api-access-s4khj\") pod \"octavia-operator-controller-manager-998648c74-cjgcs\" (UID: \"b869ba57-9812-4684-8cbb-48fc606c6aec\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.335764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47njt\" (UniqueName: \"kubernetes.io/projected/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-kube-api-access-47njt\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.375184 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnvq6\" (UniqueName: \"kubernetes.io/projected/406103b8-0898-4d8f-badd-5f32648ad242-kube-api-access-vnvq6\") pod \"swift-operator-controller-manager-5f8c65bbfc-6zgq8\" (UID: \"406103b8-0898-4d8f-badd-5f32648ad242\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.375264 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t57p\" (UniqueName: \"kubernetes.io/projected/0b693fe0-7de6-4a0c-a931-21174d92c184-kube-api-access-6t57p\") pod \"telemetry-operator-controller-manager-75c4b445dc-tfgqn\" (UID: \"0b693fe0-7de6-4a0c-a931-21174d92c184\") " pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.375312 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6sds\" (UniqueName: \"kubernetes.io/projected/b6018c51-4633-48dd-97ea-0d3d405f4e14-kube-api-access-n6sds\") pod \"placement-operator-controller-manager-78f8948974-vjb5b\" (UID: \"b6018c51-4633-48dd-97ea-0d3d405f4e14\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.375358 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgpsp\" (UniqueName: \"kubernetes.io/projected/9b79fea5-fb83-4804-987d-b61e11999829-kube-api-access-vgpsp\") pod \"ovn-operator-controller-manager-b6456fdb6-m9qpq\" (UID: \"9b79fea5-fb83-4804-987d-b61e11999829\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.388453 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.389828 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.391266 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.391386 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.395372 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-w9wv5" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.395504 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.396681 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.397759 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6sds\" (UniqueName: \"kubernetes.io/projected/b6018c51-4633-48dd-97ea-0d3d405f4e14-kube-api-access-n6sds\") pod \"placement-operator-controller-manager-78f8948974-vjb5b\" (UID: \"b6018c51-4633-48dd-97ea-0d3d405f4e14\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.399898 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.404885 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-v56cl" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.404893 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.411927 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t57p\" (UniqueName: \"kubernetes.io/projected/0b693fe0-7de6-4a0c-a931-21174d92c184-kube-api-access-6t57p\") pod \"telemetry-operator-controller-manager-75c4b445dc-tfgqn\" (UID: \"0b693fe0-7de6-4a0c-a931-21174d92c184\") " pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.424445 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.428263 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.429528 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.434856 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.435718 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.435898 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lbxjv" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.436020 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.438248 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgpsp\" (UniqueName: \"kubernetes.io/projected/9b79fea5-fb83-4804-987d-b61e11999829-kube-api-access-vgpsp\") pod \"ovn-operator-controller-manager-b6456fdb6-m9qpq\" (UID: \"9b79fea5-fb83-4804-987d-b61e11999829\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.438317 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.448545 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.449081 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnvq6\" (UniqueName: \"kubernetes.io/projected/406103b8-0898-4d8f-badd-5f32648ad242-kube-api-access-vnvq6\") pod \"swift-operator-controller-manager-5f8c65bbfc-6zgq8\" (UID: \"406103b8-0898-4d8f-badd-5f32648ad242\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.456241 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.460770 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.464741 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-5jz7h" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.464894 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.474496 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.482237 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd49g\" (UniqueName: \"kubernetes.io/projected/2c77adc3-2d13-4c70-aefa-e4c238e61547-kube-api-access-dd49g\") pod \"test-operator-controller-manager-5854674fcc-rjb6d\" (UID: \"2c77adc3-2d13-4c70-aefa-e4c238e61547\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.482421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj676\" (UniqueName: \"kubernetes.io/projected/82c72555-ba21-47f1-bfcd-ef954e788eb5-kube-api-access-lj676\") pod \"watcher-operator-controller-manager-769dc69bc-k2567\" (UID: \"82c72555-ba21-47f1-bfcd-ef954e788eb5\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.520669 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.536702 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.587641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588746 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b5qf\" (UniqueName: \"kubernetes.io/projected/123c59a4-e74e-4b51-9477-84d97ba60e36-kube-api-access-9b5qf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5j8pz\" (UID: \"123c59a4-e74e-4b51-9477-84d97ba60e36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588824 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd49g\" (UniqueName: \"kubernetes.io/projected/2c77adc3-2d13-4c70-aefa-e4c238e61547-kube-api-access-dd49g\") pod \"test-operator-controller-manager-5854674fcc-rjb6d\" (UID: \"2c77adc3-2d13-4c70-aefa-e4c238e61547\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588848 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctmb8\" (UniqueName: \"kubernetes.io/projected/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-kube-api-access-ctmb8\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588877 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588914 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.588965 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj676\" (UniqueName: \"kubernetes.io/projected/82c72555-ba21-47f1-bfcd-ef954e788eb5-kube-api-access-lj676\") pod \"watcher-operator-controller-manager-769dc69bc-k2567\" (UID: \"82c72555-ba21-47f1-bfcd-ef954e788eb5\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.590624 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.590692 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:14.590673134 +0000 UTC m=+1129.878164803 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.628941 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd49g\" (UniqueName: \"kubernetes.io/projected/2c77adc3-2d13-4c70-aefa-e4c238e61547-kube-api-access-dd49g\") pod \"test-operator-controller-manager-5854674fcc-rjb6d\" (UID: \"2c77adc3-2d13-4c70-aefa-e4c238e61547\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.674557 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj676\" (UniqueName: \"kubernetes.io/projected/82c72555-ba21-47f1-bfcd-ef954e788eb5-kube-api-access-lj676\") pod \"watcher-operator-controller-manager-769dc69bc-k2567\" (UID: \"82c72555-ba21-47f1-bfcd-ef954e788eb5\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.690159 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b5qf\" (UniqueName: \"kubernetes.io/projected/123c59a4-e74e-4b51-9477-84d97ba60e36-kube-api-access-9b5qf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5j8pz\" (UID: \"123c59a4-e74e-4b51-9477-84d97ba60e36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.690279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctmb8\" (UniqueName: \"kubernetes.io/projected/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-kube-api-access-ctmb8\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.690320 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.690354 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.690492 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.690547 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:14.190531284 +0000 UTC m=+1129.478022953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.691957 4960 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.692000 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:14.19199048 +0000 UTC m=+1129.479482149 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "metrics-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.752136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b5qf\" (UniqueName: \"kubernetes.io/projected/123c59a4-e74e-4b51-9477-84d97ba60e36-kube-api-access-9b5qf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-5j8pz\" (UID: \"123c59a4-e74e-4b51-9477-84d97ba60e36\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.765300 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.765618 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctmb8\" (UniqueName: \"kubernetes.io/projected/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-kube-api-access-ctmb8\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.792223 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.792831 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: E1201 19:58:13.792886 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert podName:6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c nodeName:}" failed. No retries permitted until 2025-12-01 19:58:14.792869013 +0000 UTC m=+1130.080360672 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" (UID: "6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.864893 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf"] Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.910858 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.956159 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:13 crc kubenswrapper[4960]: I1201 19:58:13.999808 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" event={"ID":"238b3987-76a6-401d-a33c-5c2efd1bafcf","Type":"ContainerStarted","Data":"a66a079783d7850dcbc902b8828c26be4c429c563d400257ba4b6c6b408c10de"} Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.121838 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.131369 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.199769 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.200446 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.199999 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.200753 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:15.200714897 +0000 UTC m=+1130.488206566 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.200665 4960 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.201004 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:15.200982026 +0000 UTC m=+1130.488473695 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "metrics-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: W1201 19:58:14.204708 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcce3a29f_5414_4fee_a27c_9b6286930903.slice/crio-976b5014fe709ced23e376dcdd8a2f592d25c3d1855b5a40c4ba01591fc08242 WatchSource:0}: Error finding container 976b5014fe709ced23e376dcdd8a2f592d25c3d1855b5a40c4ba01591fc08242: Status 404 returned error can't find the container with id 976b5014fe709ced23e376dcdd8a2f592d25c3d1855b5a40c4ba01591fc08242 Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.606787 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.606982 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.607071 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:16.607053546 +0000 UTC m=+1131.894545215 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.757433 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.762732 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm"] Dec 01 19:58:14 crc kubenswrapper[4960]: W1201 19:58:14.763001 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod018cf742_9a47_4355_879a_9e064f358748.slice/crio-4342192930ec6a5f5986c633fa7f36f4fdfa4ba1d78b8dcfe62c965275df90ab WatchSource:0}: Error finding container 4342192930ec6a5f5986c633fa7f36f4fdfa4ba1d78b8dcfe62c965275df90ab: Status 404 returned error can't find the container with id 4342192930ec6a5f5986c633fa7f36f4fdfa4ba1d78b8dcfe62c965275df90ab Dec 01 19:58:14 crc kubenswrapper[4960]: W1201 19:58:14.767811 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb869ba57_9812_4684_8cbb_48fc606c6aec.slice/crio-9a7c5227c7268af8132cdd5250a190707efec9b68545445ab627d7ff5f221844 WatchSource:0}: Error finding container 9a7c5227c7268af8132cdd5250a190707efec9b68545445ab627d7ff5f221844: Status 404 returned error can't find the container with id 9a7c5227c7268af8132cdd5250a190707efec9b68545445ab627d7ff5f221844 Dec 01 19:58:14 crc kubenswrapper[4960]: W1201 19:58:14.767950 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58815c45_1d27_4f86_9b6a_ffbb04c4e250.slice/crio-71f4f60285fd2d71062f0ab3cb261af50725d6b6255a79577c2dc9bf9990b63e WatchSource:0}: Error finding container 71f4f60285fd2d71062f0ab3cb261af50725d6b6255a79577c2dc9bf9990b63e: Status 404 returned error can't find the container with id 71f4f60285fd2d71062f0ab3cb261af50725d6b6255a79577c2dc9bf9990b63e Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.788007 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.810922 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.811228 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: E1201 19:58:14.811297 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert podName:6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c nodeName:}" failed. No retries permitted until 2025-12-01 19:58:16.811276177 +0000 UTC m=+1132.098767846 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" (UID: "6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.924670 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.950231 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7"] Dec 01 19:58:14 crc kubenswrapper[4960]: I1201 19:58:14.971104 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.002973 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c0036c6_9f32_4163_866d_c1f5786c5a2e.slice/crio-f14029b402d63dee99212817585cdf5795f84aabe4ab88bd1ab3d152a8fdd0ea WatchSource:0}: Error finding container f14029b402d63dee99212817585cdf5795f84aabe4ab88bd1ab3d152a8fdd0ea: Status 404 returned error can't find the container with id f14029b402d63dee99212817585cdf5795f84aabe4ab88bd1ab3d152a8fdd0ea Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.006387 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.015336 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.024263 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.031383 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b481993_f9fd_4d99_98e3_04b51c186e20.slice/crio-37a46422a8a8422db4ce6b8bcdbeab2639cca8524094a3fda4aab8d8e9394612 WatchSource:0}: Error finding container 37a46422a8a8422db4ce6b8bcdbeab2639cca8524094a3fda4aab8d8e9394612: Status 404 returned error can't find the container with id 37a46422a8a8422db4ce6b8bcdbeab2639cca8524094a3fda4aab8d8e9394612 Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.031490 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.035067 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b693fe0_7de6_4a0c_a931_21174d92c184.slice/crio-f3ecca26ae84515ffe8059cc00cb1a1ea14b3bfc2efe3b873ec77ea3abfe71ff WatchSource:0}: Error finding container f3ecca26ae84515ffe8059cc00cb1a1ea14b3bfc2efe3b873ec77ea3abfe71ff: Status 404 returned error can't find the container with id f3ecca26ae84515ffe8059cc00cb1a1ea14b3bfc2efe3b873ec77ea3abfe71ff Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.035461 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde2501b9_bdd7_46c3_9128_20b2d9d5ab88.slice/crio-27b164cba285a5a6a7073088125bb1bde2df76c61fa52fc91506c73709e2fb94 WatchSource:0}: Error finding container 27b164cba285a5a6a7073088125bb1bde2df76c61fa52fc91506c73709e2fb94: Status 404 returned error can't find the container with id 27b164cba285a5a6a7073088125bb1bde2df76c61fa52fc91506c73709e2fb94 Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.035509 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.042515 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod406103b8_0898_4d8f_badd_5f32648ad242.slice/crio-d4036d41de3de1e6052b55cfd1697ecaa1282ed9195e7750c8a766cb826dd013 WatchSource:0}: Error finding container d4036d41de3de1e6052b55cfd1697ecaa1282ed9195e7750c8a766cb826dd013: Status 404 returned error can't find the container with id d4036d41de3de1e6052b55cfd1697ecaa1282ed9195e7750c8a766cb826dd013 Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.044079 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.048901 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" event={"ID":"8b922299-0b5e-4595-bdab-1c759fd99465","Type":"ContainerStarted","Data":"6189a0fa2b790a809ac788c84f97b81648f8d63afea686e93b2c62dbba83daca"} Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.053003 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vnvq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6zgq8_openstack-operators(406103b8-0898-4d8f-badd-5f32648ad242): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.053594 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" event={"ID":"cce3a29f-5414-4fee-a27c-9b6286930903","Type":"ContainerStarted","Data":"976b5014fe709ced23e376dcdd8a2f592d25c3d1855b5a40c4ba01591fc08242"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.055234 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" event={"ID":"018cf742-9a47-4355-879a-9e064f358748","Type":"ContainerStarted","Data":"4342192930ec6a5f5986c633fa7f36f4fdfa4ba1d78b8dcfe62c965275df90ab"} Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.055331 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vnvq6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6zgq8_openstack-operators(406103b8-0898-4d8f-badd-5f32648ad242): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.056564 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" podUID="406103b8-0898-4d8f-badd-5f32648ad242" Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.058080 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" event={"ID":"0a07e8af-0ea2-4b3e-b9b7-b01f88025904","Type":"ContainerStarted","Data":"a4409ec0dbac88a3aa0b10a38f71d83d75a0c53e62b83c3fd8151143b776b69f"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.059962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" event={"ID":"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d","Type":"ContainerStarted","Data":"8caa06608496c35823a4cff44387220b2475dd8a46215d8347f067a756b2fb86"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.063703 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" event={"ID":"b869ba57-9812-4684-8cbb-48fc606c6aec","Type":"ContainerStarted","Data":"9a7c5227c7268af8132cdd5250a190707efec9b68545445ab627d7ff5f221844"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.066292 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" event={"ID":"a51010d5-7158-47ce-9428-25f5b2ca871f","Type":"ContainerStarted","Data":"b2f8d4c0f5a2721f8e7fa2d5fec2172aa4028140c25939358e74bef948b1a759"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.079953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" event={"ID":"58815c45-1d27-4f86-9b6a-ffbb04c4e250","Type":"ContainerStarted","Data":"71f4f60285fd2d71062f0ab3cb261af50725d6b6255a79577c2dc9bf9990b63e"} Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.177465 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.184896 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.194677 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.195334 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod123c59a4_e74e_4b51_9477_84d97ba60e36.slice/crio-2f0c7735e76d2e97e43dca29c90507b6ccc5f9604a1d0e3facd91435d657d86a WatchSource:0}: Error finding container 2f0c7735e76d2e97e43dca29c90507b6ccc5f9604a1d0e3facd91435d657d86a: Status 404 returned error can't find the container with id 2f0c7735e76d2e97e43dca29c90507b6ccc5f9604a1d0e3facd91435d657d86a Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.197930 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9b5qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-5j8pz_openstack-operators(123c59a4-e74e-4b51-9477-84d97ba60e36): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.199232 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podUID="123c59a4-e74e-4b51-9477-84d97ba60e36" Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.202606 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6018c51_4633_48dd_97ea_0d3d405f4e14.slice/crio-9579c10a8aa5a5d617488933f84355724f4073211df820dd6728be75dfddb1b6 WatchSource:0}: Error finding container 9579c10a8aa5a5d617488933f84355724f4073211df820dd6728be75dfddb1b6: Status 404 returned error can't find the container with id 9579c10a8aa5a5d617488933f84355724f4073211df820dd6728be75dfddb1b6 Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.205691 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n6sds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-vjb5b_openstack-operators(b6018c51-4633-48dd-97ea-0d3d405f4e14): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.208052 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n6sds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-vjb5b_openstack-operators(b6018c51-4633-48dd-97ea-0d3d405f4e14): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.209791 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" podUID="b6018c51-4633-48dd-97ea-0d3d405f4e14" Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.219807 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567"] Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.222686 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.222763 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.222894 4960 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.222981 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.222993 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:17.222966962 +0000 UTC m=+1132.510458631 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "metrics-server-cert" not found Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.223076 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:17.223057755 +0000 UTC m=+1132.510549414 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "webhook-server-cert" not found Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.223294 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82c72555_ba21_47f1_bfcd_ef954e788eb5.slice/crio-3e9bb524675871a4286e46537235c64dada2ecef20358012c4274d288abbfcd8 WatchSource:0}: Error finding container 3e9bb524675871a4286e46537235c64dada2ecef20358012c4274d288abbfcd8: Status 404 returned error can't find the container with id 3e9bb524675871a4286e46537235c64dada2ecef20358012c4274d288abbfcd8 Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.226080 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lj676,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-k2567_openstack-operators(82c72555-ba21-47f1-bfcd-ef954e788eb5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.228451 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lj676,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-k2567_openstack-operators(82c72555-ba21-47f1-bfcd-ef954e788eb5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 01 19:58:15 crc kubenswrapper[4960]: E1201 19:58:15.229744 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" podUID="82c72555-ba21-47f1-bfcd-ef954e788eb5" Dec 01 19:58:15 crc kubenswrapper[4960]: I1201 19:58:15.556454 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d"] Dec 01 19:58:15 crc kubenswrapper[4960]: W1201 19:58:15.578365 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c77adc3_2d13_4c70_aefa_e4c238e61547.slice/crio-aa47ebe99e7959724a8ffce0dcd57ad6b010f892e5a62ca3dbd80bff4ad27535 WatchSource:0}: Error finding container aa47ebe99e7959724a8ffce0dcd57ad6b010f892e5a62ca3dbd80bff4ad27535: Status 404 returned error can't find the container with id aa47ebe99e7959724a8ffce0dcd57ad6b010f892e5a62ca3dbd80bff4ad27535 Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.099499 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" event={"ID":"de2501b9-bdd7-46c3-9128-20b2d9d5ab88","Type":"ContainerStarted","Data":"27b164cba285a5a6a7073088125bb1bde2df76c61fa52fc91506c73709e2fb94"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.101758 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" event={"ID":"9b79fea5-fb83-4804-987d-b61e11999829","Type":"ContainerStarted","Data":"c330be27c06a2b5ce9fee664771aa91988206161e5621efe6313479dacdbf65d"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.104801 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" event={"ID":"3b481993-f9fd-4d99-98e3-04b51c186e20","Type":"ContainerStarted","Data":"37a46422a8a8422db4ce6b8bcdbeab2639cca8524094a3fda4aab8d8e9394612"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.111621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" event={"ID":"b6018c51-4633-48dd-97ea-0d3d405f4e14","Type":"ContainerStarted","Data":"9579c10a8aa5a5d617488933f84355724f4073211df820dd6728be75dfddb1b6"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.118556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" event={"ID":"406103b8-0898-4d8f-badd-5f32648ad242","Type":"ContainerStarted","Data":"d4036d41de3de1e6052b55cfd1697ecaa1282ed9195e7750c8a766cb826dd013"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.123926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" event={"ID":"82c72555-ba21-47f1-bfcd-ef954e788eb5","Type":"ContainerStarted","Data":"3e9bb524675871a4286e46537235c64dada2ecef20358012c4274d288abbfcd8"} Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.124543 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" podUID="b6018c51-4633-48dd-97ea-0d3d405f4e14" Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.126022 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" podUID="406103b8-0898-4d8f-badd-5f32648ad242" Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.132531 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" event={"ID":"0106714e-857b-4ade-9ee2-66fe8c12164c","Type":"ContainerStarted","Data":"8b3711773e784be663a249cf4ce84b5fbc489e5ae026310041029e49d31a78a6"} Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.133083 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" podUID="82c72555-ba21-47f1-bfcd-ef954e788eb5" Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.137461 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" event={"ID":"2c77adc3-2d13-4c70-aefa-e4c238e61547","Type":"ContainerStarted","Data":"aa47ebe99e7959724a8ffce0dcd57ad6b010f892e5a62ca3dbd80bff4ad27535"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.155346 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" event={"ID":"123c59a4-e74e-4b51-9477-84d97ba60e36","Type":"ContainerStarted","Data":"2f0c7735e76d2e97e43dca29c90507b6ccc5f9604a1d0e3facd91435d657d86a"} Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.156977 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podUID="123c59a4-e74e-4b51-9477-84d97ba60e36" Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.159266 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" event={"ID":"8c0036c6-9f32-4163-866d-c1f5786c5a2e","Type":"ContainerStarted","Data":"f14029b402d63dee99212817585cdf5795f84aabe4ab88bd1ab3d152a8fdd0ea"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.162992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" event={"ID":"0b693fe0-7de6-4a0c-a931-21174d92c184","Type":"ContainerStarted","Data":"f3ecca26ae84515ffe8059cc00cb1a1ea14b3bfc2efe3b873ec77ea3abfe71ff"} Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.679179 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.679386 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.679488 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:20.679466295 +0000 UTC m=+1135.966957964 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:16 crc kubenswrapper[4960]: I1201 19:58:16.884674 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.885030 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:16 crc kubenswrapper[4960]: E1201 19:58:16.885096 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert podName:6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c nodeName:}" failed. No retries permitted until 2025-12-01 19:58:20.885074109 +0000 UTC m=+1136.172565778 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" (UID: "6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.183502 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" podUID="406103b8-0898-4d8f-badd-5f32648ad242" Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.183930 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podUID="123c59a4-e74e-4b51-9477-84d97ba60e36" Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.184828 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" podUID="82c72555-ba21-47f1-bfcd-ef954e788eb5" Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.188030 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" podUID="b6018c51-4633-48dd-97ea-0d3d405f4e14" Dec 01 19:58:17 crc kubenswrapper[4960]: I1201 19:58:17.295211 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:17 crc kubenswrapper[4960]: I1201 19:58:17.295400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.295436 4960 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.295545 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:21.295520166 +0000 UTC m=+1136.583011825 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "metrics-server-cert" not found Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.295770 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 19:58:17 crc kubenswrapper[4960]: E1201 19:58:17.295887 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:21.295874937 +0000 UTC m=+1136.583366606 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "webhook-server-cert" not found Dec 01 19:58:20 crc kubenswrapper[4960]: I1201 19:58:20.758516 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:20 crc kubenswrapper[4960]: E1201 19:58:20.759467 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:20 crc kubenswrapper[4960]: E1201 19:58:20.759523 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:28.759505813 +0000 UTC m=+1144.046997482 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:20 crc kubenswrapper[4960]: I1201 19:58:20.963371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:20 crc kubenswrapper[4960]: E1201 19:58:20.963702 4960 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:20 crc kubenswrapper[4960]: E1201 19:58:20.963829 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert podName:6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c nodeName:}" failed. No retries permitted until 2025-12-01 19:58:28.963807628 +0000 UTC m=+1144.251299297 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" (UID: "6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 01 19:58:21 crc kubenswrapper[4960]: I1201 19:58:21.373404 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:21 crc kubenswrapper[4960]: I1201 19:58:21.373488 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:21 crc kubenswrapper[4960]: E1201 19:58:21.373809 4960 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 01 19:58:21 crc kubenswrapper[4960]: E1201 19:58:21.373881 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:29.373857302 +0000 UTC m=+1144.661348981 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "webhook-server-cert" not found Dec 01 19:58:21 crc kubenswrapper[4960]: E1201 19:58:21.374353 4960 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 01 19:58:21 crc kubenswrapper[4960]: E1201 19:58:21.374400 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs podName:d52ae874-2bed-4c1a-9c8c-a3e140971d4a nodeName:}" failed. No retries permitted until 2025-12-01 19:58:29.374387509 +0000 UTC m=+1144.661879188 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs") pod "openstack-operator-controller-manager-8466c66f58-t6nmf" (UID: "d52ae874-2bed-4c1a-9c8c-a3e140971d4a") : secret "metrics-server-cert" not found Dec 01 19:58:25 crc kubenswrapper[4960]: I1201 19:58:25.969603 4960 scope.go:117] "RemoveContainer" containerID="40020b2061215ca870d53db18936f414451e99ba1e3de6581cab2fa4979bead1" Dec 01 19:58:27 crc kubenswrapper[4960]: E1201 19:58:27.047270 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 01 19:58:27 crc kubenswrapper[4960]: E1201 19:58:27.048050 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vgpsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-m9qpq_openstack-operators(9b79fea5-fb83-4804-987d-b61e11999829): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:58:28 crc kubenswrapper[4960]: E1201 19:58:28.149272 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 01 19:58:28 crc kubenswrapper[4960]: E1201 19:58:28.149491 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vb68q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-m24v2_openstack-operators(de2501b9-bdd7-46c3-9128-20b2d9d5ab88): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:58:28 crc kubenswrapper[4960]: I1201 19:58:28.826491 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:28 crc kubenswrapper[4960]: E1201 19:58:28.826723 4960 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:28 crc kubenswrapper[4960]: E1201 19:58:28.826883 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert podName:b801b659-3926-4a9f-a36f-f752a6bd9563 nodeName:}" failed. No retries permitted until 2025-12-01 19:58:44.826857975 +0000 UTC m=+1160.114349644 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert") pod "infra-operator-controller-manager-57548d458d-pnrhx" (UID: "b801b659-3926-4a9f-a36f-f752a6bd9563") : secret "infra-operator-webhook-server-cert" not found Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.029590 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.039243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6\" (UID: \"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.060809 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5gsf7" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.068824 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.435427 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.435986 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.441275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-metrics-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.443031 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d52ae874-2bed-4c1a-9c8c-a3e140971d4a-webhook-certs\") pod \"openstack-operator-controller-manager-8466c66f58-t6nmf\" (UID: \"d52ae874-2bed-4c1a-9c8c-a3e140971d4a\") " pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.605950 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lbxjv" Dec 01 19:58:29 crc kubenswrapper[4960]: I1201 19:58:29.614161 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:30 crc kubenswrapper[4960]: E1201 19:58:30.661359 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5" Dec 01 19:58:30 crc kubenswrapper[4960]: E1201 19:58:30.661624 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:ecf7be921850bdc04697ed1b332bab39ad2a64e4e45c2a445c04f9bae6ac61b5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z5jvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-hcpbb_openstack-operators(3b481993-f9fd-4d99-98e3-04b51c186e20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:58:31 crc kubenswrapper[4960]: E1201 19:58:31.449865 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 01 19:58:31 crc kubenswrapper[4960]: E1201 19:58:31.450415 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-559bk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-ld68p_openstack-operators(0106714e-857b-4ade-9ee2-66fe8c12164c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:58:31 crc kubenswrapper[4960]: I1201 19:58:31.469548 4960 scope.go:117] "RemoveContainer" containerID="d17e91ff1d4a802f157bc45dc86d0f146998befcc97ccc681ce3cc4928116154" Dec 01 19:58:33 crc kubenswrapper[4960]: I1201 19:58:33.874141 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf"] Dec 01 19:58:43 crc kubenswrapper[4960]: I1201 19:58:43.009418 4960 scope.go:117] "RemoveContainer" containerID="82023d1ce19f77ff1045907ea78a3e9f6dd280032bc9ff0bd575d25d3eaf05ef" Dec 01 19:58:43 crc kubenswrapper[4960]: W1201 19:58:43.024659 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd52ae874_2bed_4c1a_9c8c_a3e140971d4a.slice/crio-e13c66f9943b4a54a19fb12ce817a8d22a7faa3201422306561aa2fc548b63ad WatchSource:0}: Error finding container e13c66f9943b4a54a19fb12ce817a8d22a7faa3201422306561aa2fc548b63ad: Status 404 returned error can't find the container with id e13c66f9943b4a54a19fb12ce817a8d22a7faa3201422306561aa2fc548b63ad Dec 01 19:58:43 crc kubenswrapper[4960]: I1201 19:58:43.421803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" event={"ID":"d52ae874-2bed-4c1a-9c8c-a3e140971d4a","Type":"ContainerStarted","Data":"e13c66f9943b4a54a19fb12ce817a8d22a7faa3201422306561aa2fc548b63ad"} Dec 01 19:58:43 crc kubenswrapper[4960]: I1201 19:58:43.461379 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6"] Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.552225 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.552387 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9b5qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-5j8pz_openstack-operators(123c59a4-e74e-4b51-9477-84d97ba60e36): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.553961 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podUID="123c59a4-e74e-4b51-9477-84d97ba60e36" Dec 01 19:58:43 crc kubenswrapper[4960]: W1201 19:58:43.603720 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6994eaf1_2e28_45bb_8ae3_7c5bd2b51e9c.slice/crio-005958c9b1b6592a2be614fa336efb01d02cc58da9b9f9a815dd466a849c3cb6 WatchSource:0}: Error finding container 005958c9b1b6592a2be614fa336efb01d02cc58da9b9f9a815dd466a849c3cb6: Status 404 returned error can't find the container with id 005958c9b1b6592a2be614fa336efb01d02cc58da9b9f9a815dd466a849c3cb6 Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.692486 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.692652 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vb68q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-m24v2_openstack-operators(de2501b9-bdd7-46c3-9128-20b2d9d5ab88): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.693923 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" podUID="de2501b9-bdd7-46c3-9128-20b2d9d5ab88" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.702083 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.702280 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z5jvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-hcpbb_openstack-operators(3b481993-f9fd-4d99-98e3-04b51c186e20): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 01 19:58:43 crc kubenswrapper[4960]: E1201 19:58:43.703449 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" podUID="3b481993-f9fd-4d99-98e3-04b51c186e20" Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.435434 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" event={"ID":"0a07e8af-0ea2-4b3e-b9b7-b01f88025904","Type":"ContainerStarted","Data":"a4eec783010fae637beda6d4392fc8036d7661226560c29eec1a584d7fcdab00"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.437867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" event={"ID":"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c","Type":"ContainerStarted","Data":"005958c9b1b6592a2be614fa336efb01d02cc58da9b9f9a815dd466a849c3cb6"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.441682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" event={"ID":"b869ba57-9812-4684-8cbb-48fc606c6aec","Type":"ContainerStarted","Data":"bcfdfdd99c5637f1912b1e38b8ccc7f5c026c1b95ee4826a3cef868c67a60bb6"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.444007 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" event={"ID":"58815c45-1d27-4f86-9b6a-ffbb04c4e250","Type":"ContainerStarted","Data":"f11a466f227c29f32b9f99cd91ff813b4467f75365e27d91fe8ee889969bffcb"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.445638 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" event={"ID":"8c0036c6-9f32-4163-866d-c1f5786c5a2e","Type":"ContainerStarted","Data":"4ec58a7a636de2d575661f9d3044223c25eafc6e646edce663df3bf594da5bb5"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.448002 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" event={"ID":"018cf742-9a47-4355-879a-9e064f358748","Type":"ContainerStarted","Data":"3be7487cb088bcb363952d18223a3087caaefb6fd0eaa29e9f797e165f5243f6"} Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.896868 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.904975 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b801b659-3926-4a9f-a36f-f752a6bd9563-cert\") pod \"infra-operator-controller-manager-57548d458d-pnrhx\" (UID: \"b801b659-3926-4a9f-a36f-f752a6bd9563\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:44 crc kubenswrapper[4960]: I1201 19:58:44.998748 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-hlx5s" Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.003417 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.479691 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" event={"ID":"cce3a29f-5414-4fee-a27c-9b6286930903","Type":"ContainerStarted","Data":"7448a442305560c170d43e3e134043781003205bbf792a6bfdb3a8c7755a0957"} Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.506645 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" event={"ID":"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d","Type":"ContainerStarted","Data":"73c34a93988dd57655590552eb3f69451e6ba93e19009fc12b2649e9a548697a"} Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.511980 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" event={"ID":"a51010d5-7158-47ce-9428-25f5b2ca871f","Type":"ContainerStarted","Data":"06a3af703e54e5ec4ed28f2a92f10ab7cdbf7d3850030e20aa1ed2409a2b6cd3"} Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.523945 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" event={"ID":"238b3987-76a6-401d-a33c-5c2efd1bafcf","Type":"ContainerStarted","Data":"86cf64730e4e209f6b8982b06b2d004732d68db97a30c836edf3861ec9df4bc9"} Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.560358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" event={"ID":"8b922299-0b5e-4595-bdab-1c759fd99465","Type":"ContainerStarted","Data":"075e0e07213306c6373d35e2faa45db856c4e000e01eed9328d7fc94779e0e30"} Dec 01 19:58:45 crc kubenswrapper[4960]: I1201 19:58:45.579396 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" event={"ID":"2c77adc3-2d13-4c70-aefa-e4c238e61547","Type":"ContainerStarted","Data":"bf4d7eada46f861f7d600683b0ce001bab1be288d43878109ad71cf883d473df"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.253238 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx"] Dec 01 19:58:46 crc kubenswrapper[4960]: E1201 19:58:46.404546 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 01 19:58:46 crc kubenswrapper[4960]: E1201 19:58:46.404722 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vgpsp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-m9qpq_openstack-operators(9b79fea5-fb83-4804-987d-b61e11999829): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 01 19:58:46 crc kubenswrapper[4960]: E1201 19:58:46.405900 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" podUID="9b79fea5-fb83-4804-987d-b61e11999829" Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.606196 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" event={"ID":"406103b8-0898-4d8f-badd-5f32648ad242","Type":"ContainerStarted","Data":"b112e41f2751ce9423b0bc3d3c5441119880358f038082bc3ac6e7aa189441aa"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.610876 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" event={"ID":"82c72555-ba21-47f1-bfcd-ef954e788eb5","Type":"ContainerStarted","Data":"7ade4d65e4fd7ab47180d1ef4f675c1f33136c831df095d3523096d6b5cb7e39"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.612319 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" event={"ID":"d52ae874-2bed-4c1a-9c8c-a3e140971d4a","Type":"ContainerStarted","Data":"3663d3fbe7fd7db00b0eb8176f47182ffd947f3cb0ba26139c72d194e4e7fb62"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.613406 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.614250 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" event={"ID":"b801b659-3926-4a9f-a36f-f752a6bd9563","Type":"ContainerStarted","Data":"312e4985fdd9ef2300b911d791ba98a2edf99cf98f08386db57b2cba442cdeb3"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.615787 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" event={"ID":"0b693fe0-7de6-4a0c-a931-21174d92c184","Type":"ContainerStarted","Data":"0cd9cf6bb825bab7d0cd3e2d15a00e07dfb6f76954dabd7c49822d4763810651"} Dec 01 19:58:46 crc kubenswrapper[4960]: I1201 19:58:46.652944 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" podStartSLOduration=33.652927845 podStartE2EDuration="33.652927845s" podCreationTimestamp="2025-12-01 19:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 19:58:46.649629143 +0000 UTC m=+1161.937120822" watchObservedRunningTime="2025-12-01 19:58:46.652927845 +0000 UTC m=+1161.940419514" Dec 01 19:58:48 crc kubenswrapper[4960]: I1201 19:58:48.633084 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" event={"ID":"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c","Type":"ContainerStarted","Data":"76f3ddc0baf782d55c7d180d850e41260437d8d82648bc0f61e5ac3c4c3c34f2"} Dec 01 19:58:48 crc kubenswrapper[4960]: I1201 19:58:48.638642 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" event={"ID":"3b481993-f9fd-4d99-98e3-04b51c186e20","Type":"ContainerStarted","Data":"dccfbad8171bafe4996ddf7ed64e51a01fc561eb5c1568c5b860847fe37770c0"} Dec 01 19:58:48 crc kubenswrapper[4960]: I1201 19:58:48.640565 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" event={"ID":"b6018c51-4633-48dd-97ea-0d3d405f4e14","Type":"ContainerStarted","Data":"79642322481c9b9134a8c2ef2a64b5ef0d76303821313f58f3e90efa0c630a66"} Dec 01 19:58:48 crc kubenswrapper[4960]: I1201 19:58:48.642218 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" event={"ID":"de2501b9-bdd7-46c3-9128-20b2d9d5ab88","Type":"ContainerStarted","Data":"a888d168d5a7b099d9dcc7d2ae8868c58f0851fbbb48cfd358a7f0d7e348809d"} Dec 01 19:58:48 crc kubenswrapper[4960]: I1201 19:58:48.672049 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" podStartSLOduration=2.903626465 podStartE2EDuration="35.672025043s" podCreationTimestamp="2025-12-01 19:58:13 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.582359678 +0000 UTC m=+1130.869851337" lastFinishedPulling="2025-12-01 19:58:48.350758236 +0000 UTC m=+1163.638249915" observedRunningTime="2025-12-01 19:58:48.664463828 +0000 UTC m=+1163.951955497" watchObservedRunningTime="2025-12-01 19:58:48.672025043 +0000 UTC m=+1163.959516712" Dec 01 19:58:49 crc kubenswrapper[4960]: E1201 19:58:49.145945 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" podUID="0106714e-857b-4ade-9ee2-66fe8c12164c" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.666459 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" event={"ID":"0a07e8af-0ea2-4b3e-b9b7-b01f88025904","Type":"ContainerStarted","Data":"710f939ecc98c323afee0a86c019a97b5fd13a722a7881708248444170d7b6ec"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.666663 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.669322 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.669739 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" event={"ID":"3b481993-f9fd-4d99-98e3-04b51c186e20","Type":"ContainerStarted","Data":"ca421445084005d8904c69e7f710b2876bfa5aba9059609fa33af1ac668e936a"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.669812 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.678066 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" event={"ID":"a51010d5-7158-47ce-9428-25f5b2ca871f","Type":"ContainerStarted","Data":"7a012daa5442d19012b8b60634e2d141913651dc8bbf51a4f7f957884888c5fc"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.678171 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.684614 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" event={"ID":"0b693fe0-7de6-4a0c-a931-21174d92c184","Type":"ContainerStarted","Data":"e00eb23f28767513a0428b674e58f39841d37e97d48ca61d4df343bb8b137533"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.684743 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.693229 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-6mqj7" podStartSLOduration=4.266565172 podStartE2EDuration="37.693215995s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.987099134 +0000 UTC m=+1130.274590793" lastFinishedPulling="2025-12-01 19:58:48.413749947 +0000 UTC m=+1163.701241616" observedRunningTime="2025-12-01 19:58:49.686051312 +0000 UTC m=+1164.973542981" watchObservedRunningTime="2025-12-01 19:58:49.693215995 +0000 UTC m=+1164.980707664" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.702035 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" event={"ID":"58815c45-1d27-4f86-9b6a-ffbb04c4e250","Type":"ContainerStarted","Data":"16a7913c2dcd2314b5ddd3ece952777dfb1b470e805f4063c548b288e954ce9e"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.702256 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.705611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" event={"ID":"8b922299-0b5e-4595-bdab-1c759fd99465","Type":"ContainerStarted","Data":"e054a840ad7bef666f8809f0326d5326c3f60200f3f7705fce9a1e0341a01465"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.705673 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.706789 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.710924 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" event={"ID":"9b79fea5-fb83-4804-987d-b61e11999829","Type":"ContainerStarted","Data":"a7b38318218c6b507a216c639239b90a8476857baf3aff17ca98f30ecf27e082"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.710958 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" event={"ID":"9b79fea5-fb83-4804-987d-b61e11999829","Type":"ContainerStarted","Data":"06ea7517082c97b2c6198262d937fe37bdfe328c84aa44460edbc4d82a26f557"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.711152 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.715812 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" podStartSLOduration=3.400672469 podStartE2EDuration="36.715799458s" podCreationTimestamp="2025-12-01 19:58:13 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.052476631 +0000 UTC m=+1130.339968300" lastFinishedPulling="2025-12-01 19:58:48.36760362 +0000 UTC m=+1163.655095289" observedRunningTime="2025-12-01 19:58:49.713305031 +0000 UTC m=+1165.000796700" watchObservedRunningTime="2025-12-01 19:58:49.715799458 +0000 UTC m=+1165.003291127" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.719372 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" event={"ID":"82c72555-ba21-47f1-bfcd-ef954e788eb5","Type":"ContainerStarted","Data":"d141bc8138ee4759fbcded134f4fbb8e79c9ea52616261203ee220315d37ceb1"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.719495 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.726228 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" event={"ID":"0106714e-857b-4ade-9ee2-66fe8c12164c","Type":"ContainerStarted","Data":"49b19d78c5ba9e6ccfb12780018ef3200343f4ad175ed4b9032aa6e3f1abef5f"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.735830 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" podStartSLOduration=6.398815205 podStartE2EDuration="37.735821502s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.05243772 +0000 UTC m=+1130.339929389" lastFinishedPulling="2025-12-01 19:58:46.389444017 +0000 UTC m=+1161.676935686" observedRunningTime="2025-12-01 19:58:49.733369506 +0000 UTC m=+1165.020861165" watchObservedRunningTime="2025-12-01 19:58:49.735821502 +0000 UTC m=+1165.023313171" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.738525 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" event={"ID":"6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c","Type":"ContainerStarted","Data":"f572c720a4fe9512fc109446c49a81c5e56fe2bca5c71c2fc542f5b9b90e3ac0"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.739233 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.755391 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" event={"ID":"b869ba57-9812-4684-8cbb-48fc606c6aec","Type":"ContainerStarted","Data":"ee57466c25570ccb304bbb86c80fdc8a963a24107848556b8751107ee4d4d505"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.755638 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.757374 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" event={"ID":"238b3987-76a6-401d-a33c-5c2efd1bafcf","Type":"ContainerStarted","Data":"cff2071329caa231019c8f8e73dc8c0118af42a34e8be9480832b52c3249abb5"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.757853 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.761588 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" event={"ID":"de2501b9-bdd7-46c3-9128-20b2d9d5ab88","Type":"ContainerStarted","Data":"91f5808c4224a415cb53ce0c7898099be74be8efa4e81491203e65d3b5e61425"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.761953 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.763994 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" event={"ID":"2c77adc3-2d13-4c70-aefa-e4c238e61547","Type":"ContainerStarted","Data":"4a4709f39067265fe4cd52dbb3c3e261e2e5faa4702a340271d77b357f06d845"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.764363 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.764606 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.768330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" event={"ID":"8c0036c6-9f32-4163-866d-c1f5786c5a2e","Type":"ContainerStarted","Data":"02a07afea1125b3d15d66ee5052e97c17badf249a8f9a6587110a7fd961774cf"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.768685 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.770936 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" podStartSLOduration=3.603404403 podStartE2EDuration="37.770921096s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.215049294 +0000 UTC m=+1129.502540963" lastFinishedPulling="2025-12-01 19:58:48.382565987 +0000 UTC m=+1163.670057656" observedRunningTime="2025-12-01 19:58:49.768405428 +0000 UTC m=+1165.055897097" watchObservedRunningTime="2025-12-01 19:58:49.770921096 +0000 UTC m=+1165.058412765" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.771717 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.774009 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" event={"ID":"dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d","Type":"ContainerStarted","Data":"1e432253c4c373a53f3960b4111de6454947a516c2e1ea775cfcb05b63f3bd7d"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.774202 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.782253 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" event={"ID":"b6018c51-4633-48dd-97ea-0d3d405f4e14","Type":"ContainerStarted","Data":"2be90dd5c39f082a4e48b1eb959bf02ff1d4fbf5b6ccb8019a984a51e7551230"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.782489 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.799265 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" event={"ID":"406103b8-0898-4d8f-badd-5f32648ad242","Type":"ContainerStarted","Data":"ee9dd9b40daaafa3bc7fd02b00d88cde378b6b0739f8453c1a7a8d0ded78e7bf"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.799961 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.815878 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" event={"ID":"cce3a29f-5414-4fee-a27c-9b6286930903","Type":"ContainerStarted","Data":"0ddac71a0c468ea673513a655a1ef1777fc1b429b03197d3bc02818cccf87fbd"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.816598 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.826395 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" event={"ID":"018cf742-9a47-4355-879a-9e064f358748","Type":"ContainerStarted","Data":"689b26c954181541abddaefd236a14e3e296aac06974d91197876ec459ac6705"} Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.827187 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.831014 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" podStartSLOduration=4.851964518 podStartE2EDuration="37.831004517s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.186900879 +0000 UTC m=+1130.474392558" lastFinishedPulling="2025-12-01 19:58:48.165940868 +0000 UTC m=+1163.453432557" observedRunningTime="2025-12-01 19:58:49.830404438 +0000 UTC m=+1165.117896107" watchObservedRunningTime="2025-12-01 19:58:49.831004517 +0000 UTC m=+1165.118496186" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.839474 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.903250 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" podStartSLOduration=4.721874316 podStartE2EDuration="37.903232297s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.022339363 +0000 UTC m=+1130.309831032" lastFinishedPulling="2025-12-01 19:58:48.203697334 +0000 UTC m=+1163.491189013" observedRunningTime="2025-12-01 19:58:49.900236514 +0000 UTC m=+1165.187728183" watchObservedRunningTime="2025-12-01 19:58:49.903232297 +0000 UTC m=+1165.190723966" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.908668 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" podStartSLOduration=3.962001044 podStartE2EDuration="36.908656726s" podCreationTimestamp="2025-12-01 19:58:13 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.225927844 +0000 UTC m=+1130.513419513" lastFinishedPulling="2025-12-01 19:58:48.172583516 +0000 UTC m=+1163.460075195" observedRunningTime="2025-12-01 19:58:49.879409976 +0000 UTC m=+1165.166901645" watchObservedRunningTime="2025-12-01 19:58:49.908656726 +0000 UTC m=+1165.196148385" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.953959 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" podStartSLOduration=33.39475586 podStartE2EDuration="37.953942047s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:43.613367748 +0000 UTC m=+1158.900859417" lastFinishedPulling="2025-12-01 19:58:48.172553935 +0000 UTC m=+1163.460045604" observedRunningTime="2025-12-01 19:58:49.951586534 +0000 UTC m=+1165.239078203" watchObservedRunningTime="2025-12-01 19:58:49.953942047 +0000 UTC m=+1165.241433726" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.955318 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" podStartSLOduration=4.523874597 podStartE2EDuration="37.95531082s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.95070173 +0000 UTC m=+1130.238193419" lastFinishedPulling="2025-12-01 19:58:48.382137973 +0000 UTC m=+1163.669629642" observedRunningTime="2025-12-01 19:58:49.924296314 +0000 UTC m=+1165.211787973" watchObservedRunningTime="2025-12-01 19:58:49.95531082 +0000 UTC m=+1165.242802489" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.970801 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" podStartSLOduration=3.594388372 podStartE2EDuration="37.970782862s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:13.974365776 +0000 UTC m=+1129.261857445" lastFinishedPulling="2025-12-01 19:58:48.350760266 +0000 UTC m=+1163.638251935" observedRunningTime="2025-12-01 19:58:49.968425938 +0000 UTC m=+1165.255917607" watchObservedRunningTime="2025-12-01 19:58:49.970782862 +0000 UTC m=+1165.258274531" Dec 01 19:58:49 crc kubenswrapper[4960]: I1201 19:58:49.993733 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-cjgcs" podStartSLOduration=4.43667243 podStartE2EDuration="37.993719256s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.797303212 +0000 UTC m=+1130.084794881" lastFinishedPulling="2025-12-01 19:58:48.354350038 +0000 UTC m=+1163.641841707" observedRunningTime="2025-12-01 19:58:49.99127365 +0000 UTC m=+1165.278765319" watchObservedRunningTime="2025-12-01 19:58:49.993719256 +0000 UTC m=+1165.281210915" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.047080 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-zggh7" podStartSLOduration=4.453283878 podStartE2EDuration="38.047065718s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.791610075 +0000 UTC m=+1130.079101734" lastFinishedPulling="2025-12-01 19:58:48.385391905 +0000 UTC m=+1163.672883574" observedRunningTime="2025-12-01 19:58:50.046063527 +0000 UTC m=+1165.333555196" watchObservedRunningTime="2025-12-01 19:58:50.047065718 +0000 UTC m=+1165.334557377" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.048371 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-q8bv7" podStartSLOduration=4.89663153 podStartE2EDuration="38.048364439s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.051841771 +0000 UTC m=+1130.339333440" lastFinishedPulling="2025-12-01 19:58:48.20357469 +0000 UTC m=+1163.491066349" observedRunningTime="2025-12-01 19:58:50.020088298 +0000 UTC m=+1165.307579967" watchObservedRunningTime="2025-12-01 19:58:50.048364439 +0000 UTC m=+1165.335856108" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.066349 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" podStartSLOduration=7.672639356 podStartE2EDuration="38.066330698s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.20557119 +0000 UTC m=+1130.493062859" lastFinishedPulling="2025-12-01 19:58:45.599262532 +0000 UTC m=+1160.886754201" observedRunningTime="2025-12-01 19:58:50.064821002 +0000 UTC m=+1165.352312671" watchObservedRunningTime="2025-12-01 19:58:50.066330698 +0000 UTC m=+1165.353822367" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.086446 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" podStartSLOduration=6.749975763 podStartE2EDuration="38.086427034s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.05180587 +0000 UTC m=+1130.339297539" lastFinishedPulling="2025-12-01 19:58:46.388257141 +0000 UTC m=+1161.675748810" observedRunningTime="2025-12-01 19:58:50.079419206 +0000 UTC m=+1165.366910875" watchObservedRunningTime="2025-12-01 19:58:50.086427034 +0000 UTC m=+1165.373918703" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.128622 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-pbxhm" podStartSLOduration=4.527308674 podStartE2EDuration="38.128605128s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.766308396 +0000 UTC m=+1130.053800065" lastFinishedPulling="2025-12-01 19:58:48.36760485 +0000 UTC m=+1163.655096519" observedRunningTime="2025-12-01 19:58:50.104895769 +0000 UTC m=+1165.392387438" watchObservedRunningTime="2025-12-01 19:58:50.128605128 +0000 UTC m=+1165.416096797" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.129957 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" podStartSLOduration=4.799408361 podStartE2EDuration="38.12994612s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.052641436 +0000 UTC m=+1130.340133105" lastFinishedPulling="2025-12-01 19:58:48.383179185 +0000 UTC m=+1163.670670864" observedRunningTime="2025-12-01 19:58:50.123962794 +0000 UTC m=+1165.411454463" watchObservedRunningTime="2025-12-01 19:58:50.12994612 +0000 UTC m=+1165.417437789" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.835029 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" event={"ID":"0106714e-857b-4ade-9ee2-66fe8c12164c","Type":"ContainerStarted","Data":"d1afb55f000baf473c7a22f44b220c43f4352b78efa77417e33ac68dc6de2ca0"} Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.835202 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.837206 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" event={"ID":"b801b659-3926-4a9f-a36f-f752a6bd9563","Type":"ContainerStarted","Data":"801dac65988df0d5241c4def1cdc641864556dea369991c1c2cd7d4d3ef1dd12"} Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.842161 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-k2567" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.842208 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-bsfbn" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.842228 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-6tbkt" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.842246 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.842944 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rjb6d" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.843089 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-8l7sf" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.843304 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-75c4b445dc-tfgqn" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.844888 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-sfsvg" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.856850 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" podStartSLOduration=3.324142694 podStartE2EDuration="38.856832994s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.044324237 +0000 UTC m=+1130.331815906" lastFinishedPulling="2025-12-01 19:58:50.577014537 +0000 UTC m=+1165.864506206" observedRunningTime="2025-12-01 19:58:50.855371898 +0000 UTC m=+1166.142863567" watchObservedRunningTime="2025-12-01 19:58:50.856832994 +0000 UTC m=+1166.144324663" Dec 01 19:58:50 crc kubenswrapper[4960]: I1201 19:58:50.864570 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-p5qfh" podStartSLOduration=4.692878652 podStartE2EDuration="38.864546524s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:14.21140338 +0000 UTC m=+1129.498895049" lastFinishedPulling="2025-12-01 19:58:48.383071242 +0000 UTC m=+1163.670562921" observedRunningTime="2025-12-01 19:58:50.142991837 +0000 UTC m=+1165.430483516" watchObservedRunningTime="2025-12-01 19:58:50.864546524 +0000 UTC m=+1166.152038193" Dec 01 19:58:51 crc kubenswrapper[4960]: I1201 19:58:51.850675 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" event={"ID":"b801b659-3926-4a9f-a36f-f752a6bd9563","Type":"ContainerStarted","Data":"d6176fcb5ba577185ee752792f065e2bf196fb98e639d76f027f615a6e40a914"} Dec 01 19:58:51 crc kubenswrapper[4960]: I1201 19:58:51.855483 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6zgq8" Dec 01 19:58:51 crc kubenswrapper[4960]: I1201 19:58:51.884274 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" podStartSLOduration=35.687612993 podStartE2EDuration="39.884245235s" podCreationTimestamp="2025-12-01 19:58:12 +0000 UTC" firstStartedPulling="2025-12-01 19:58:46.394244917 +0000 UTC m=+1161.681736576" lastFinishedPulling="2025-12-01 19:58:50.590877149 +0000 UTC m=+1165.878368818" observedRunningTime="2025-12-01 19:58:51.878176376 +0000 UTC m=+1167.165668085" watchObservedRunningTime="2025-12-01 19:58:51.884245235 +0000 UTC m=+1167.171736934" Dec 01 19:58:52 crc kubenswrapper[4960]: I1201 19:58:52.862339 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:53 crc kubenswrapper[4960]: I1201 19:58:53.393211 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-hcpbb" Dec 01 19:58:53 crc kubenswrapper[4960]: I1201 19:58:53.407987 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-m24v2" Dec 01 19:58:53 crc kubenswrapper[4960]: I1201 19:58:53.479803 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-vjb5b" Dec 01 19:58:53 crc kubenswrapper[4960]: I1201 19:58:53.524198 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-m9qpq" Dec 01 19:58:55 crc kubenswrapper[4960]: I1201 19:58:55.013421 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-pnrhx" Dec 01 19:58:58 crc kubenswrapper[4960]: E1201 19:58:58.326620 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podUID="123c59a4-e74e-4b51-9477-84d97ba60e36" Dec 01 19:58:59 crc kubenswrapper[4960]: I1201 19:58:59.075434 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6" Dec 01 19:58:59 crc kubenswrapper[4960]: I1201 19:58:59.622512 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8466c66f58-t6nmf" Dec 01 19:59:03 crc kubenswrapper[4960]: I1201 19:59:03.441023 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-ld68p" Dec 01 19:59:12 crc kubenswrapper[4960]: I1201 19:59:12.039697 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" event={"ID":"123c59a4-e74e-4b51-9477-84d97ba60e36","Type":"ContainerStarted","Data":"c99ae4ace8afca5f2b18a42bf9ca21e1d69b8367d139e150261910b03aabe995"} Dec 01 19:59:12 crc kubenswrapper[4960]: I1201 19:59:12.058223 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-5j8pz" podStartSLOduration=3.531673917 podStartE2EDuration="59.058203255s" podCreationTimestamp="2025-12-01 19:58:13 +0000 UTC" firstStartedPulling="2025-12-01 19:58:15.197760397 +0000 UTC m=+1130.485252066" lastFinishedPulling="2025-12-01 19:59:10.724289735 +0000 UTC m=+1186.011781404" observedRunningTime="2025-12-01 19:59:12.05481439 +0000 UTC m=+1187.342306069" watchObservedRunningTime="2025-12-01 19:59:12.058203255 +0000 UTC m=+1187.345694934" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.558738 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.565758 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.569917 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-x8zvd" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.569924 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.570006 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.571045 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.577015 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.713725 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.715250 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5x7q\" (UniqueName: \"kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.715438 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.715727 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.717992 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.742467 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.817133 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.817184 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.817210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.817251 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5x7q\" (UniqueName: \"kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.817307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbtbt\" (UniqueName: \"kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.818165 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.841187 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5x7q\" (UniqueName: \"kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q\") pod \"dnsmasq-dns-675f4bcbfc-jmlv2\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.886408 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.918647 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.918704 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.918778 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbtbt\" (UniqueName: \"kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.919685 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.919725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:29 crc kubenswrapper[4960]: I1201 19:59:29.947384 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbtbt\" (UniqueName: \"kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt\") pod \"dnsmasq-dns-78dd6ddcc-qjw7l\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:30 crc kubenswrapper[4960]: I1201 19:59:30.029686 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 19:59:30 crc kubenswrapper[4960]: I1201 19:59:30.125605 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 19:59:30 crc kubenswrapper[4960]: W1201 19:59:30.133015 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod332d7faf_04ea_49e5_8dfc_764444f702cb.slice/crio-6b1f2646ba792b911c0d29e46a72b6face63fb159dff77e38cdc5d982fb8af83 WatchSource:0}: Error finding container 6b1f2646ba792b911c0d29e46a72b6face63fb159dff77e38cdc5d982fb8af83: Status 404 returned error can't find the container with id 6b1f2646ba792b911c0d29e46a72b6face63fb159dff77e38cdc5d982fb8af83 Dec 01 19:59:30 crc kubenswrapper[4960]: I1201 19:59:30.216981 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" event={"ID":"332d7faf-04ea-49e5-8dfc-764444f702cb","Type":"ContainerStarted","Data":"6b1f2646ba792b911c0d29e46a72b6face63fb159dff77e38cdc5d982fb8af83"} Dec 01 19:59:30 crc kubenswrapper[4960]: I1201 19:59:30.483018 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 19:59:30 crc kubenswrapper[4960]: W1201 19:59:30.483560 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85bc78b4_f077_43b0_8a5a_5f56301a6af6.slice/crio-840f99f03aca75df83ff96940c0b123d04211951d7440ec627b996ecf1e96ee8 WatchSource:0}: Error finding container 840f99f03aca75df83ff96940c0b123d04211951d7440ec627b996ecf1e96ee8: Status 404 returned error can't find the container with id 840f99f03aca75df83ff96940c0b123d04211951d7440ec627b996ecf1e96ee8 Dec 01 19:59:31 crc kubenswrapper[4960]: I1201 19:59:31.228282 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" event={"ID":"85bc78b4-f077-43b0-8a5a-5f56301a6af6","Type":"ContainerStarted","Data":"840f99f03aca75df83ff96940c0b123d04211951d7440ec627b996ecf1e96ee8"} Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.520253 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.551270 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.569270 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.588689 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.663893 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.663963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68pht\" (UniqueName: \"kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.664021 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.765809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.765892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.765949 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68pht\" (UniqueName: \"kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.767166 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.767700 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.795011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68pht\" (UniqueName: \"kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht\") pod \"dnsmasq-dns-666b6646f7-zztmq\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.848801 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.872147 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.873783 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.883192 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.899222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.968302 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm5m6\" (UniqueName: \"kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.968419 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:32.968460 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.070278 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.070625 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.070692 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm5m6\" (UniqueName: \"kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.071448 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.071681 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.094235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm5m6\" (UniqueName: \"kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6\") pod \"dnsmasq-dns-57d769cc4f-qmxlv\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.204032 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.704058 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.706714 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.710069 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.710068 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.712188 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.712487 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-l27tk" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.713900 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.714072 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.718176 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.721072 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.881316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.881372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.881421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.881452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.881717 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882179 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882293 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882346 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrpz6\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882641 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882668 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.882788 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.967469 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.978331 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 19:59:33 crc kubenswrapper[4960]: W1201 19:59:33.978462 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod795f6715_b4c6_410f_868c_eeb7139913d1.slice/crio-388745035a30e18a1cd72a2b7efad43e05ecc28a80d8ee04538645765ae77c7d WatchSource:0}: Error finding container 388745035a30e18a1cd72a2b7efad43e05ecc28a80d8ee04538645765ae77c7d: Status 404 returned error can't find the container with id 388745035a30e18a1cd72a2b7efad43e05ecc28a80d8ee04538645765ae77c7d Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.985821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.985887 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.985954 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrpz6\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.985988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986024 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986049 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986078 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986130 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986161 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986238 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.986311 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.988394 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.989532 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.990159 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.990688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.991193 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.993665 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.994080 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d5d5199b574f54ea5db19b4cd834168e52f09a99fc0718b4c879e6f4546f0a6d/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.994461 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:33 crc kubenswrapper[4960]: I1201 19:59:33.994538 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:33.996616 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.002778 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.000536 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.008805 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.008879 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.009048 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.009321 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-57t9l" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.009447 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.009650 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.012103 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.021822 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.054871 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.056245 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrpz6\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6\") pod \"rabbitmq-server-0\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.056751 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088498 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2shq5\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088574 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088619 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088671 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088737 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088756 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088777 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088799 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.088853 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193532 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193636 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193672 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193715 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193760 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2shq5\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193814 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.193896 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.194023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.194050 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.195965 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.196306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.197096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.200408 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.203505 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.203931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.206993 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.207164 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.207209 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0004aa7191d203af0e3c5da42d773fcf606e2a73631a908716312932b18ed0b8/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.216619 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.217386 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.227086 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2shq5\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.247460 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.259873 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" event={"ID":"b484313f-1265-4ac3-8d73-5dbe81926118","Type":"ContainerStarted","Data":"3d4fd8b6f9fb2dcbfc1030328e31ac117aa6bb4178cde8d112baa68930dd4100"} Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.266470 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" event={"ID":"795f6715-b4c6-410f-868c-eeb7139913d1","Type":"ContainerStarted","Data":"388745035a30e18a1cd72a2b7efad43e05ecc28a80d8ee04538645765ae77c7d"} Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.349556 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.466174 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 19:59:34 crc kubenswrapper[4960]: I1201 19:59:34.843002 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 19:59:34 crc kubenswrapper[4960]: W1201 19:59:34.849466 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2b84514_def8_482d_a4a3_8b71808fcdcc.slice/crio-41f45c327a130a59df07f3fce5873893aae2aaca7e32e3d0a7c14822ea43c4be WatchSource:0}: Error finding container 41f45c327a130a59df07f3fce5873893aae2aaca7e32e3d0a7c14822ea43c4be: Status 404 returned error can't find the container with id 41f45c327a130a59df07f3fce5873893aae2aaca7e32e3d0a7c14822ea43c4be Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.006570 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 19:59:35 crc kubenswrapper[4960]: W1201 19:59:35.016686 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87ca6976_ad61_40c2_97a4_56ca46df7448.slice/crio-bb8ea855b9728ce71426fad40c31513b5bd8e1c5e75002956a4f5345eb236daf WatchSource:0}: Error finding container bb8ea855b9728ce71426fad40c31513b5bd8e1c5e75002956a4f5345eb236daf: Status 404 returned error can't find the container with id bb8ea855b9728ce71426fad40c31513b5bd8e1c5e75002956a4f5345eb236daf Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.278032 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerStarted","Data":"bb8ea855b9728ce71426fad40c31513b5bd8e1c5e75002956a4f5345eb236daf"} Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.280365 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerStarted","Data":"41f45c327a130a59df07f3fce5873893aae2aaca7e32e3d0a7c14822ea43c4be"} Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.363680 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.365205 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.368615 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.368800 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lqwdk" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.368913 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.369028 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.378728 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.388578 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530096 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530524 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kolla-config\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530573 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530596 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6fdd\" (UniqueName: \"kubernetes.io/projected/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kube-api-access-c6fdd\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530613 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-default\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530680 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.530702 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632604 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632686 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kolla-config\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632728 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632744 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6fdd\" (UniqueName: \"kubernetes.io/projected/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kube-api-access-c6fdd\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632766 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-default\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632801 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.632859 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.633483 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.633785 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kolla-config\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.634819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-config-data-default\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.634973 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.636035 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.636065 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/125e246f4bb097af8e6ffec5afb4769badc7eaa601e21d55b2122d431e565718/globalmount\"" pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.638387 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.647024 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.650006 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6fdd\" (UniqueName: \"kubernetes.io/projected/f9b50b6f-1af6-4283-8994-1cdc2b9e6c02-kube-api-access-c6fdd\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.694826 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-67cfe74a-27f7-4e2e-9ad8-9c6486fb1fa7\") pod \"openstack-galera-0\" (UID: \"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02\") " pod="openstack/openstack-galera-0" Dec 01 19:59:35 crc kubenswrapper[4960]: I1201 19:59:35.997222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.517496 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.950607 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.952071 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.954630 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-mk9mg" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.957391 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.957575 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.957707 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 01 19:59:36 crc kubenswrapper[4960]: I1201 19:59:36.964551 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.061914 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.062094 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fsgt\" (UniqueName: \"kubernetes.io/projected/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kube-api-access-2fsgt\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063239 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063378 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063503 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063541 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.063580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165349 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165662 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fsgt\" (UniqueName: \"kubernetes.io/projected/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kube-api-access-2fsgt\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165728 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165756 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165798 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165818 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.165868 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.166322 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.166849 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.168324 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.169410 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.172386 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.172418 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b53d75724c7344f053c509d80398620b345c23d3bdf47132ec79c67ccb0b1b60/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.179223 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.186079 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.197059 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fsgt\" (UniqueName: \"kubernetes.io/projected/c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2-kube-api-access-2fsgt\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.281401 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9ec85f1-62a7-48d7-996b-cb925cf5d23b\") pod \"openstack-cell1-galera-0\" (UID: \"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2\") " pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.342714 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02","Type":"ContainerStarted","Data":"f0ecae55087975ea5c35f36bb40a783291871cfe8c8a65c2496f094a594c0370"} Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.483635 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.484671 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.490452 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.490452 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tdwr4" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.492051 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.505043 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.588622 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.594573 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.594632 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kolla-config\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.594662 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.594692 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-config-data\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.594726 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp82c\" (UniqueName: \"kubernetes.io/projected/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kube-api-access-sp82c\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.696428 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.696488 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kolla-config\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.696546 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.696576 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-config-data\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.696609 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp82c\" (UniqueName: \"kubernetes.io/projected/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kube-api-access-sp82c\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.700754 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-config-data\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.701834 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kolla-config\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.708062 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.711387 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.716091 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp82c\" (UniqueName: \"kubernetes.io/projected/6befffd7-b8c4-4c6c-900c-f6e1728f2a9e-kube-api-access-sp82c\") pod \"memcached-0\" (UID: \"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e\") " pod="openstack/memcached-0" Dec 01 19:59:37 crc kubenswrapper[4960]: I1201 19:59:37.811667 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 01 19:59:38 crc kubenswrapper[4960]: I1201 19:59:38.293780 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 01 19:59:38 crc kubenswrapper[4960]: W1201 19:59:38.308049 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2b2a9ff_2995_4e32_bc7d_49b7fc3b86d2.slice/crio-dbfe359fa3dfcff88b644e33fd04d844a8fdda50ac5e60de813bb066ada42302 WatchSource:0}: Error finding container dbfe359fa3dfcff88b644e33fd04d844a8fdda50ac5e60de813bb066ada42302: Status 404 returned error can't find the container with id dbfe359fa3dfcff88b644e33fd04d844a8fdda50ac5e60de813bb066ada42302 Dec 01 19:59:38 crc kubenswrapper[4960]: I1201 19:59:38.365092 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2","Type":"ContainerStarted","Data":"dbfe359fa3dfcff88b644e33fd04d844a8fdda50ac5e60de813bb066ada42302"} Dec 01 19:59:38 crc kubenswrapper[4960]: I1201 19:59:38.445234 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 01 19:59:38 crc kubenswrapper[4960]: W1201 19:59:38.460445 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6befffd7_b8c4_4c6c_900c_f6e1728f2a9e.slice/crio-af07ae199c23fd63a94735e7de854fd82cb1eeea8498daca40b09a81f8bd4f4f WatchSource:0}: Error finding container af07ae199c23fd63a94735e7de854fd82cb1eeea8498daca40b09a81f8bd4f4f: Status 404 returned error can't find the container with id af07ae199c23fd63a94735e7de854fd82cb1eeea8498daca40b09a81f8bd4f4f Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.150709 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.152215 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.155841 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6qtzz" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.177828 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.237870 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-755hk\" (UniqueName: \"kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk\") pod \"kube-state-metrics-0\" (UID: \"7e7b2075-1317-4327-a8d9-4ebda5a68344\") " pod="openstack/kube-state-metrics-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.340275 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-755hk\" (UniqueName: \"kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk\") pod \"kube-state-metrics-0\" (UID: \"7e7b2075-1317-4327-a8d9-4ebda5a68344\") " pod="openstack/kube-state-metrics-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.386073 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e","Type":"ContainerStarted","Data":"af07ae199c23fd63a94735e7de854fd82cb1eeea8498daca40b09a81f8bd4f4f"} Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.396300 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-755hk\" (UniqueName: \"kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk\") pod \"kube-state-metrics-0\" (UID: \"7e7b2075-1317-4327-a8d9-4ebda5a68344\") " pod="openstack/kube-state-metrics-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.493667 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.978057 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.980811 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:39 crc kubenswrapper[4960]: I1201 19:59:39.986751 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:39.994730 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:39.994883 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-pzrdb" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:39.995025 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:39.995049 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:39.995107 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.065878 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.065965 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.066038 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.066084 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.066133 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.066152 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.066187 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlfrx\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-kube-api-access-zlfrx\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170512 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170571 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170628 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170668 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170731 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.170769 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlfrx\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-kube-api-access-zlfrx\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.176428 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.177373 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.201914 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.202101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/39fbd706-99ef-4b5d-9d7a-e84ee337c443-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.206437 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.206788 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/39fbd706-99ef-4b5d-9d7a-e84ee337c443-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.210670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlfrx\" (UniqueName: \"kubernetes.io/projected/39fbd706-99ef-4b5d-9d7a-e84ee337c443-kube-api-access-zlfrx\") pod \"alertmanager-metric-storage-0\" (UID: \"39fbd706-99ef-4b5d-9d7a-e84ee337c443\") " pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.321697 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.508876 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.510886 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.519250 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.519464 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.519621 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.521237 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-lkllm" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.521529 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.521658 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.537914 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.595627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.595699 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmtcn\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.595727 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.595969 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.596060 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.596193 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.596222 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.596254 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.697604 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.697904 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.697969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.697990 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.698015 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.698083 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmtcn\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.698102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.698157 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.698936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.703689 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.703727 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/aa508db444fabf42cff0f8d4f7fdb19d89c1265d06a9375978c8e8544e26865b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.709392 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.714441 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.720837 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.721266 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.736677 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.736932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmtcn\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.752392 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:40 crc kubenswrapper[4960]: I1201 19:59:40.834689 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.501986 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.519301 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.519754 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.523078 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.523399 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.523890 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.524549 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.531559 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-f8dw6" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.571749 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.571835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.571875 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.571903 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.571951 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.577464 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.577498 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrq4d\" (UniqueName: \"kubernetes.io/projected/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-kube-api-access-hrq4d\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.577525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.639687 4960 scope.go:117] "RemoveContainer" containerID="1016e2782ebac5b8c32f8a5042c2e901e149b677005c278b4a8612728e0d0022" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.678997 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679082 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679105 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrq4d\" (UniqueName: \"kubernetes.io/projected/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-kube-api-access-hrq4d\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679152 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679214 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679272 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679313 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.679371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.680671 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.681153 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.686828 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.686881 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b75cde1bdca518731ebc83a5547272919390266c12b9cad55cf48164af54c37c/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.692962 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.696061 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.696678 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.697376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.722571 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrq4d\" (UniqueName: \"kubernetes.io/projected/4a7eb168-4c2f-46c0-9292-de8e5c4383a5-kube-api-access-hrq4d\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.759677 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68030cc2-7b9b-4dbc-acd4-de326596c35c\") pod \"ovsdbserver-nb-0\" (UID: \"4a7eb168-4c2f-46c0-9292-de8e5c4383a5\") " pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.850444 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.873753 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.873811 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.888415 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nrf7z"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.889761 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.893907 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.894041 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dgrd8" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.894358 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.924179 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrf7z"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.934405 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-9n6l4"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.936716 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.944489 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9n6l4"] Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986599 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986732 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-lib\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986769 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9389e08d-0203-4f96-8278-fa1780a4201b-scripts\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986800 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7q5d\" (UniqueName: \"kubernetes.io/projected/9389e08d-0203-4f96-8278-fa1780a4201b-kube-api-access-w7q5d\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986825 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-etc-ovs\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986866 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e16436-45f5-4392-b524-b86f1cc5c0b2-scripts\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986915 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-log\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986947 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xjq\" (UniqueName: \"kubernetes.io/projected/87e16436-45f5-4392-b524-b86f1cc5c0b2-kube-api-access-s2xjq\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986966 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-run\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.986990 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-log-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.987022 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-combined-ca-bundle\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:43 crc kubenswrapper[4960]: I1201 19:59:43.987058 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-ovn-controller-tls-certs\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088380 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9389e08d-0203-4f96-8278-fa1780a4201b-scripts\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088442 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7q5d\" (UniqueName: \"kubernetes.io/projected/9389e08d-0203-4f96-8278-fa1780a4201b-kube-api-access-w7q5d\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088469 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-etc-ovs\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088509 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e16436-45f5-4392-b524-b86f1cc5c0b2-scripts\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088545 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-log\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088567 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xjq\" (UniqueName: \"kubernetes.io/projected/87e16436-45f5-4392-b524-b86f1cc5c0b2-kube-api-access-s2xjq\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088582 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-run\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088606 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-log-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088634 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-combined-ca-bundle\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088658 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-ovn-controller-tls-certs\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088692 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088727 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.088745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-lib\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089089 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-etc-ovs\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-log-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089325 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-log\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089516 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-lib\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/87e16436-45f5-4392-b524-b86f1cc5c0b2-var-run-ovn\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.089612 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9389e08d-0203-4f96-8278-fa1780a4201b-var-run\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.092378 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9389e08d-0203-4f96-8278-fa1780a4201b-scripts\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.094280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e16436-45f5-4392-b524-b86f1cc5c0b2-scripts\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.099319 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-ovn-controller-tls-certs\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.114426 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xjq\" (UniqueName: \"kubernetes.io/projected/87e16436-45f5-4392-b524-b86f1cc5c0b2-kube-api-access-s2xjq\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.124028 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7q5d\" (UniqueName: \"kubernetes.io/projected/9389e08d-0203-4f96-8278-fa1780a4201b-kube-api-access-w7q5d\") pod \"ovn-controller-ovs-9n6l4\" (UID: \"9389e08d-0203-4f96-8278-fa1780a4201b\") " pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.132605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e16436-45f5-4392-b524-b86f1cc5c0b2-combined-ca-bundle\") pod \"ovn-controller-nrf7z\" (UID: \"87e16436-45f5-4392-b524-b86f1cc5c0b2\") " pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.234796 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z" Dec 01 19:59:44 crc kubenswrapper[4960]: I1201 19:59:44.259239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 19:59:46 crc kubenswrapper[4960]: I1201 19:59:46.649669 4960 scope.go:117] "RemoveContainer" containerID="9977357e08d68d2be3da0a3f6b1aff32c248eed26d0eda8802531e5d43acc1bc" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.019469 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.029345 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.032737 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.033058 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dc7tr" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.034060 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.036691 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.048546 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168100 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-config\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168167 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168231 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168272 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168302 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168366 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.168396 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvg4p\" (UniqueName: \"kubernetes.io/projected/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-kube-api-access-fvg4p\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270261 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270310 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270423 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270685 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvg4p\" (UniqueName: \"kubernetes.io/projected/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-kube-api-access-fvg4p\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270744 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-config\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270847 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.270919 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.271936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.272177 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-config\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.272524 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.278468 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.278546 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/527ce550ab76bd98b13c05c0daf9e637a5f97564b0f439318e7db621c3d9b898/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.281694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.281920 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.284136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.291379 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvg4p\" (UniqueName: \"kubernetes.io/projected/1c6f94f2-a256-4e77-b6d1-39706b7c6ddc-kube-api-access-fvg4p\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.312782 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-93b42492-3bdd-4312-b6cd-7c39c71c0c38\") pod \"ovsdbserver-sb-0\" (UID: \"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc\") " pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.389498 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.780260 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj"] Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.781533 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.788269 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-config" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.788438 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-dockercfg-tk6wg" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.788538 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca-bundle" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.788676 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-grpc" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.789880 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-distributor-http" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.803958 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj"] Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.884142 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.884190 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.884234 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.884404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.884450 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9blv\" (UniqueName: \"kubernetes.io/projected/2e566703-0d27-4306-bd42-7a91da29c736-kube-api-access-d9blv\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.967279 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz"] Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.972429 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.975391 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-loki-s3" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.975574 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-http" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.980917 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-querier-grpc" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.989942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.989982 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.990037 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.990088 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9blv\" (UniqueName: \"kubernetes.io/projected/2e566703-0d27-4306-bd42-7a91da29c736-kube-api-access-d9blv\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.990125 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:48 crc kubenswrapper[4960]: I1201 19:59:48.992669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:48.997884 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e566703-0d27-4306-bd42-7a91da29c736-config\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:48.998395 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-http\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-http\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.000430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/2e566703-0d27-4306-bd42-7a91da29c736-cloudkitty-lokistack-distributor-grpc\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.024805 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9blv\" (UniqueName: \"kubernetes.io/projected/2e566703-0d27-4306-bd42-7a91da29c736-kube-api-access-d9blv\") pod \"cloudkitty-lokistack-distributor-56cd74f89f-jxqfj\" (UID: \"2e566703-0d27-4306-bd42-7a91da29c736\") " pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.046819 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092323 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092687 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crxkc\" (UniqueName: \"kubernetes.io/projected/60517c3e-f430-453c-b6d4-3cdf124e100b-kube-api-access-crxkc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-config\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092729 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.092744 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.107418 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.111190 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.112682 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.115678 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-grpc" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.115899 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-query-frontend-http" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.123755 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.189169 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.190627 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196048 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196267 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzvg\" (UniqueName: \"kubernetes.io/projected/2d2cf081-f13a-411b-bf24-5302a947dd9a-kube-api-access-fxzvg\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196386 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crxkc\" (UniqueName: \"kubernetes.io/projected/60517c3e-f430-453c-b6d4-3cdf124e100b-kube-api-access-crxkc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196472 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-config\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196560 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196847 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.196979 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.205736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.205862 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.205974 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.205298 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199783 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-ca" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.202099 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.200997 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60517c3e-f430-453c-b6d4-3cdf124e100b-config\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199836 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199859 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-http" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199927 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199965 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-dockercfg-p68fw" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.199984 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"cloudkitty-lokistack-gateway-ca-bundle" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.200015 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-gateway-client-http" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.210802 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-http\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-http\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.210898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-querier-grpc\" (UniqueName: \"kubernetes.io/secret/60517c3e-f430-453c-b6d4-3cdf124e100b-cloudkitty-lokistack-querier-grpc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.220437 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crxkc\" (UniqueName: \"kubernetes.io/projected/60517c3e-f430-453c-b6d4-3cdf124e100b-kube-api-access-crxkc\") pod \"cloudkitty-lokistack-querier-548665d79b-gd6lz\" (UID: \"60517c3e-f430-453c-b6d4-3cdf124e100b\") " pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.230554 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.243202 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.245191 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.259493 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307172 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307207 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307238 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph9qq\" (UniqueName: \"kubernetes.io/projected/5cc4eeaa-88b7-4439-a39f-5e889845181e-kube-api-access-ph9qq\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307367 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307424 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307472 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307515 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307577 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307623 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307681 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzvg\" (UniqueName: \"kubernetes.io/projected/2d2cf081-f13a-411b-bf24-5302a947dd9a-kube-api-access-fxzvg\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307697 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307753 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.307773 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.308691 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.309955 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d2cf081-f13a-411b-bf24-5302a947dd9a-config\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.320124 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-http\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.329421 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/2d2cf081-f13a-411b-bf24-5302a947dd9a-cloudkitty-lokistack-query-frontend-grpc\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.329896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzvg\" (UniqueName: \"kubernetes.io/projected/2d2cf081-f13a-411b-bf24-5302a947dd9a-kube-api-access-fxzvg\") pod \"cloudkitty-lokistack-query-frontend-779849886d-hf7g4\" (UID: \"2d2cf081-f13a-411b-bf24-5302a947dd9a\") " pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.379272 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409556 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409604 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409632 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxgd7\" (UniqueName: \"kubernetes.io/projected/a8b0ac50-f375-427e-b103-0bfd272201ea-kube-api-access-qxgd7\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409657 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409686 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409723 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409774 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409811 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph9qq\" (UniqueName: \"kubernetes.io/projected/5cc4eeaa-88b7-4439-a39f-5e889845181e-kube-api-access-ph9qq\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409834 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409860 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409883 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409909 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409960 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.409986 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: E1201 19:59:49.410125 4960 secret.go:188] Couldn't get secret openstack/cloudkitty-lokistack-gateway-http: secret "cloudkitty-lokistack-gateway-http" not found Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.410144 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.410539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.411064 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.411222 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.411470 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: E1201 19:59:49.418476 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret podName:5cc4eeaa-88b7-4439-a39f-5e889845181e nodeName:}" failed. No retries permitted until 2025-12-01 19:59:49.918439087 +0000 UTC m=+1225.205930846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret") pod "cloudkitty-lokistack-gateway-76cc998948-mpddm" (UID: "5cc4eeaa-88b7-4439-a39f-5e889845181e") : secret "cloudkitty-lokistack-gateway-http" not found Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.418827 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.418903 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.422459 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.426856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.427291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph9qq\" (UniqueName: \"kubernetes.io/projected/5cc4eeaa-88b7-4439-a39f-5e889845181e-kube-api-access-ph9qq\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.443338 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512415 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512507 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512539 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512659 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512716 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxgd7\" (UniqueName: \"kubernetes.io/projected/a8b0ac50-f375-427e-b103-0bfd272201ea-kube-api-access-qxgd7\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512825 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512872 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.512913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.516209 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.516878 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.516926 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-lokistack-gateway-client-http\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.517084 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-lokistack-gateway\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.517268 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.517447 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-cloudkitty-ca-bundle\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.518319 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/a8b0ac50-f375-427e-b103-0bfd272201ea-rbac\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.520602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/a8b0ac50-f375-427e-b103-0bfd272201ea-tenants\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.531681 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxgd7\" (UniqueName: \"kubernetes.io/projected/a8b0ac50-f375-427e-b103-0bfd272201ea-kube-api-access-qxgd7\") pod \"cloudkitty-lokistack-gateway-76cc998948-xw4bs\" (UID: \"a8b0ac50-f375-427e-b103-0bfd272201ea\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.569837 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.919074 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.923684 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/5cc4eeaa-88b7-4439-a39f-5e889845181e-tls-secret\") pod \"cloudkitty-lokistack-gateway-76cc998948-mpddm\" (UID: \"5cc4eeaa-88b7-4439-a39f-5e889845181e\") " pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.935584 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.936762 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.939891 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-http" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.940045 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-ingester-grpc" Dec 01 19:59:49 crc kubenswrapper[4960]: I1201 19:59:49.947327 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.012082 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.013729 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.015310 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-grpc" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.015519 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-compactor-http" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.020900 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.020939 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.020966 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.021007 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.021032 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4sw4\" (UniqueName: \"kubernetes.io/projected/969241f0-8883-4570-94d4-fdb20ee094cd-kube-api-access-m4sw4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.021065 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.021101 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.021161 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.024910 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.109005 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.110767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.114324 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-grpc" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.114515 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-lokistack-index-gateway-http" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.119779 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.124760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.124875 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.124906 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.124947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.124979 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125130 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125198 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125405 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4sw4\" (UniqueName: \"kubernetes.io/projected/969241f0-8883-4570-94d4-fdb20ee094cd-kube-api-access-m4sw4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125451 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125512 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125599 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125622 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125675 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p5xf\" (UniqueName: \"kubernetes.io/projected/8e8a9388-578d-4d41-a2cf-d2ad6e344288-kube-api-access-6p5xf\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125703 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125784 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.125915 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.129493 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.130658 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/969241f0-8883-4570-94d4-fdb20ee094cd-config\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.131819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-grpc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.134767 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ingester-http\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-lokistack-ingester-http\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.136140 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/969241f0-8883-4570-94d4-fdb20ee094cd-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.148568 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.148654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.148663 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4sw4\" (UniqueName: \"kubernetes.io/projected/969241f0-8883-4570-94d4-fdb20ee094cd-kube-api-access-m4sw4\") pod \"cloudkitty-lokistack-ingester-0\" (UID: \"969241f0-8883-4570-94d4-fdb20ee094cd\") " pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.152759 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227528 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227647 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227692 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227778 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227837 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.227935 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228016 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p5xf\" (UniqueName: \"kubernetes.io/projected/8e8a9388-578d-4d41-a2cf-d2ad6e344288-kube-api-access-6p5xf\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228042 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228102 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228182 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmlqz\" (UniqueName: \"kubernetes.io/projected/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-kube-api-access-rmlqz\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228252 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228388 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.228822 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.229794 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-config\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.230162 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.232615 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-grpc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.233359 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-compactor-http\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-lokistack-compactor-http\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.235322 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/8e8a9388-578d-4d41-a2cf-d2ad6e344288-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.247742 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p5xf\" (UniqueName: \"kubernetes.io/projected/8e8a9388-578d-4d41-a2cf-d2ad6e344288-kube-api-access-6p5xf\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.264908 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.277518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"cloudkitty-lokistack-compactor-0\" (UID: \"8e8a9388-578d-4d41-a2cf-d2ad6e344288\") " pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.329917 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.329992 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330029 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330099 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330185 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330225 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmlqz\" (UniqueName: \"kubernetes.io/projected/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-kube-api-access-rmlqz\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330270 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.330579 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.331465 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-ca-bundle\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.331659 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.331680 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-config\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.333787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-loki-s3\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-loki-s3\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.334631 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-grpc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.336357 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloudkitty-lokistack-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-cloudkitty-lokistack-index-gateway-http\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.346162 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmlqz\" (UniqueName: \"kubernetes.io/projected/eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38-kube-api-access-rmlqz\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.351447 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"cloudkitty-lokistack-index-gateway-0\" (UID: \"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38\") " pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:50 crc kubenswrapper[4960]: I1201 19:59:50.430797 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 19:59:57 crc kubenswrapper[4960]: E1201 19:59:57.150847 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 01 19:59:57 crc kubenswrapper[4960]: E1201 19:59:57.151434 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n559h658h67h674h66fh87h5fch5ffhbh576h597h558h5b6h55h58bhdch7h674h685h68h589hb4hbdh559hc9h646h577h66bh54fh678h84h5d7q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp82c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(6befffd7-b8c4-4c6c-900c-f6e1728f2a9e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:59:57 crc kubenswrapper[4960]: E1201 19:59:57.152622 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="6befffd7-b8c4-4c6c-900c-f6e1728f2a9e" Dec 01 19:59:57 crc kubenswrapper[4960]: I1201 19:59:57.566428 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrf7z"] Dec 01 19:59:57 crc kubenswrapper[4960]: E1201 19:59:57.618875 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="6befffd7-b8c4-4c6c-900c-f6e1728f2a9e" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.054936 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.055217 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrpz6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(c2b84514-def8-482d-a4a3-8b71808fcdcc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.056457 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.067502 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.067861 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2shq5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(87ca6976-ad61-40c2-97a4-56ca46df7448): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.069026 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.639712 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" Dec 01 19:59:59 crc kubenswrapper[4960]: E1201 19:59:59.639994 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.154046 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg"] Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.155902 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.158368 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.159010 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.165272 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg"] Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.222387 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.222553 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc6vl\" (UniqueName: \"kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.222583 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.323870 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.323960 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc6vl\" (UniqueName: \"kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.323980 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.324825 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.335740 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.356317 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc6vl\" (UniqueName: \"kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl\") pod \"collect-profiles-29410320-lfrqg\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: I1201 20:00:00.481536 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.826212 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.826398 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c6fdd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(f9b50b6f-1af6-4283-8994-1cdc2b9e6c02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.827895 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="f9b50b6f-1af6-4283-8994-1cdc2b9e6c02" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.834033 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.834268 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2fsgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:00 crc kubenswrapper[4960]: E1201 20:00:00.835413 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2" Dec 01 20:00:01 crc kubenswrapper[4960]: E1201 20:00:01.668785 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="f9b50b6f-1af6-4283-8994-1cdc2b9e6c02" Dec 01 20:00:01 crc kubenswrapper[4960]: E1201 20:00:01.669086 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2" Dec 01 20:00:04 crc kubenswrapper[4960]: I1201 20:00:04.485223 4960 scope.go:117] "RemoveContainer" containerID="caa3ee1e590d89841e8b868a9a2c2aeeb638bf3d383bfd4cb3248f2450eece7f" Dec 01 20:00:04 crc kubenswrapper[4960]: I1201 20:00:04.694923 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrf7z" event={"ID":"87e16436-45f5-4392-b524-b86f1cc5c0b2","Type":"ContainerStarted","Data":"8f7c5d125759a96a93ad8624767469f9f7b2a131943193ef1942237adb372c6a"} Dec 01 20:00:04 crc kubenswrapper[4960]: I1201 20:00:04.955997 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:00:05 crc kubenswrapper[4960]: I1201 20:00:05.039923 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-index-gateway-0"] Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.690660 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.691244 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xbtbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-qjw7l_openstack(85bc78b4-f077-43b0-8a5a-5f56301a6af6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.692763 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" podUID="85bc78b4-f077-43b0-8a5a-5f56301a6af6" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.694256 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.694416 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-68pht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-zztmq_openstack(795f6715-b4c6-410f-868c-eeb7139913d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.696458 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.740327 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.749343 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.749501 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mm5m6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-qmxlv_openstack(b484313f-1265-4ac3-8d73-5dbe81926118): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.750677 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.786204 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.786460 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5x7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jmlv2_openstack(332d7faf-04ea-49e5-8dfc-764444f702cb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:00:06 crc kubenswrapper[4960]: E1201 20:00:06.788500 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" podUID="332d7faf-04ea-49e5-8dfc-764444f702cb" Dec 01 20:00:06 crc kubenswrapper[4960]: I1201 20:00:06.802031 4960 scope.go:117] "RemoveContainer" containerID="dc0909dadcf8a0a4740499f0968c8686d2759bee18731fbbfd14ec561c369b7d" Dec 01 20:00:06 crc kubenswrapper[4960]: I1201 20:00:06.919204 4960 scope.go:117] "RemoveContainer" containerID="c7bc951e1de7d985bc88fab794668e2baeac2ca48cfd8c351edb32e6bda7c34f" Dec 01 20:00:06 crc kubenswrapper[4960]: I1201 20:00:06.961754 4960 scope.go:117] "RemoveContainer" containerID="463654f6c57cac8b5cf868a0c7dd34808c1c6f63b99ad29e3848167b1a08b0e8" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.340523 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.465739 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc\") pod \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.465836 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbtbt\" (UniqueName: \"kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt\") pod \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.465991 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config\") pod \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\" (UID: \"85bc78b4-f077-43b0-8a5a-5f56301a6af6\") " Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.467103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "85bc78b4-f077-43b0-8a5a-5f56301a6af6" (UID: "85bc78b4-f077-43b0-8a5a-5f56301a6af6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.467757 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config" (OuterVolumeSpecName: "config") pod "85bc78b4-f077-43b0-8a5a-5f56301a6af6" (UID: "85bc78b4-f077-43b0-8a5a-5f56301a6af6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.475383 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt" (OuterVolumeSpecName: "kube-api-access-xbtbt") pod "85bc78b4-f077-43b0-8a5a-5f56301a6af6" (UID: "85bc78b4-f077-43b0-8a5a-5f56301a6af6"). InnerVolumeSpecName "kube-api-access-xbtbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.568788 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.568834 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85bc78b4-f077-43b0-8a5a-5f56301a6af6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.568847 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbtbt\" (UniqueName: \"kubernetes.io/projected/85bc78b4-f077-43b0-8a5a-5f56301a6af6-kube-api-access-xbtbt\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.597912 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-9n6l4"] Dec 01 20:00:07 crc kubenswrapper[4960]: W1201 20:00:07.605758 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9389e08d_0203_4f96_8278_fa1780a4201b.slice/crio-5e5720a5e62569d2ce0413c1ab016e07144ac8507985da5f5da46d74d272f1b5 WatchSource:0}: Error finding container 5e5720a5e62569d2ce0413c1ab016e07144ac8507985da5f5da46d74d272f1b5: Status 404 returned error can't find the container with id 5e5720a5e62569d2ce0413c1ab016e07144ac8507985da5f5da46d74d272f1b5 Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.703254 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.718537 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz"] Dec 01 20:00:07 crc kubenswrapper[4960]: W1201 20:00:07.736929 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60517c3e_f430_453c_b6d4_3cdf124e100b.slice/crio-d648524567732f8f41f9735bf7cb84264f5634e30b2a138d94589820af03a16c WatchSource:0}: Error finding container d648524567732f8f41f9735bf7cb84264f5634e30b2a138d94589820af03a16c: Status 404 returned error can't find the container with id d648524567732f8f41f9735bf7cb84264f5634e30b2a138d94589820af03a16c Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.738519 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9n6l4" event={"ID":"9389e08d-0203-4f96-8278-fa1780a4201b","Type":"ContainerStarted","Data":"5e5720a5e62569d2ce0413c1ab016e07144ac8507985da5f5da46d74d272f1b5"} Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.738856 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.740788 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" event={"ID":"2e566703-0d27-4306-bd42-7a91da29c736","Type":"ContainerStarted","Data":"9bb49d85adfe1e46cffdc2dde699d3606a4f2405a0a2e056061ece49f4bd4e46"} Dec 01 20:00:07 crc kubenswrapper[4960]: W1201 20:00:07.747382 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39fbd706_99ef_4b5d_9d7a_e84ee337c443.slice/crio-e0be3b2ed1fe8c9679d308b4ed46cd1ff2f49c5f0601966a18385697ae0a6425 WatchSource:0}: Error finding container e0be3b2ed1fe8c9679d308b4ed46cd1ff2f49c5f0601966a18385697ae0a6425: Status 404 returned error can't find the container with id e0be3b2ed1fe8c9679d308b4ed46cd1ff2f49c5f0601966a18385697ae0a6425 Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.746812 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-compactor-0"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.748023 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38","Type":"ContainerStarted","Data":"15749f25e25710acda03a5de32359a8136fcf69a2bd5c53a7dd57e46f6530020"} Dec 01 20:00:07 crc kubenswrapper[4960]: W1201 20:00:07.748974 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d2cf081_f13a_411b_bf24_5302a947dd9a.slice/crio-51d53e9a6ab0220c1f41abd50ab9bff8ce951970753b1134d143048b13d6909a WatchSource:0}: Error finding container 51d53e9a6ab0220c1f41abd50ab9bff8ce951970753b1134d143048b13d6909a: Status 404 returned error can't find the container with id 51d53e9a6ab0220c1f41abd50ab9bff8ce951970753b1134d143048b13d6909a Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.751796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" event={"ID":"85bc78b4-f077-43b0-8a5a-5f56301a6af6","Type":"ContainerDied","Data":"840f99f03aca75df83ff96940c0b123d04211951d7440ec627b996ecf1e96ee8"} Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.751936 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qjw7l" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.754071 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.768911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerStarted","Data":"02b7df7b8aaa82675031f62eb345d619b02e6ad789da191811b50770022c806c"} Dec 01 20:00:07 crc kubenswrapper[4960]: E1201 20:00:07.775386 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.851004 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.859541 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qjw7l"] Dec 01 20:00:07 crc kubenswrapper[4960]: I1201 20:00:07.873965 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.011486 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-ingester-0"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.021738 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.028782 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs"] Dec 01 20:00:08 crc kubenswrapper[4960]: W1201 20:00:08.036248 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc107ba8b_ef9e_458f_a16f_f28f0625a85f.slice/crio-875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc WatchSource:0}: Error finding container 875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc: Status 404 returned error can't find the container with id 875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc Dec 01 20:00:08 crc kubenswrapper[4960]: W1201 20:00:08.038612 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cc4eeaa_88b7_4439_a39f_5e889845181e.slice/crio-a3853ddfc0b1612088f40e65300776863d775a61d36675b4d57f1f1541327b42 WatchSource:0}: Error finding container a3853ddfc0b1612088f40e65300776863d775a61d36675b4d57f1f1541327b42: Status 404 returned error can't find the container with id a3853ddfc0b1612088f40e65300776863d775a61d36675b4d57f1f1541327b42 Dec 01 20:00:08 crc kubenswrapper[4960]: W1201 20:00:08.039589 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e7b2075_1317_4327_a8d9_4ebda5a68344.slice/crio-7b63b8911a6cc3f1de2643e90ff71a126d94fa08c9ca751441271681db8c061a WatchSource:0}: Error finding container 7b63b8911a6cc3f1de2643e90ff71a126d94fa08c9ca751441271681db8c061a: Status 404 returned error can't find the container with id 7b63b8911a6cc3f1de2643e90ff71a126d94fa08c9ca751441271681db8c061a Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.039670 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.054136 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.078321 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 01 20:00:08 crc kubenswrapper[4960]: W1201 20:00:08.085814 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c6f94f2_a256_4e77_b6d1_39706b7c6ddc.slice/crio-4064dc0440a4f2bfe7839354526eb059ccffe45f58c80ceb5401667fe760855c WatchSource:0}: Error finding container 4064dc0440a4f2bfe7839354526eb059ccffe45f58c80ceb5401667fe760855c: Status 404 returned error can't find the container with id 4064dc0440a4f2bfe7839354526eb059ccffe45f58c80ceb5401667fe760855c Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.131821 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.293408 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5x7q\" (UniqueName: \"kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q\") pod \"332d7faf-04ea-49e5-8dfc-764444f702cb\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.293483 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config\") pod \"332d7faf-04ea-49e5-8dfc-764444f702cb\" (UID: \"332d7faf-04ea-49e5-8dfc-764444f702cb\") " Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.294181 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config" (OuterVolumeSpecName: "config") pod "332d7faf-04ea-49e5-8dfc-764444f702cb" (UID: "332d7faf-04ea-49e5-8dfc-764444f702cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.306449 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q" (OuterVolumeSpecName: "kube-api-access-z5x7q") pod "332d7faf-04ea-49e5-8dfc-764444f702cb" (UID: "332d7faf-04ea-49e5-8dfc-764444f702cb"). InnerVolumeSpecName "kube-api-access-z5x7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.396444 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5x7q\" (UniqueName: \"kubernetes.io/projected/332d7faf-04ea-49e5-8dfc-764444f702cb-kube-api-access-z5x7q\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.396505 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/332d7faf-04ea-49e5-8dfc-764444f702cb-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.778364 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"969241f0-8883-4570-94d4-fdb20ee094cd","Type":"ContainerStarted","Data":"e590a3249c9772c199ae955d6549e2e3055f1cc1524dd6d58bddd7756b0a3ccd"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.779873 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a7eb168-4c2f-46c0-9292-de8e5c4383a5","Type":"ContainerStarted","Data":"5fb665648c67a03fe428ec92a6c5f7b96cf016d7abadfed5887344f90ce426d4"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.781426 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" event={"ID":"a8b0ac50-f375-427e-b103-0bfd272201ea","Type":"ContainerStarted","Data":"4810978528c9415fce9aafb96260fbbec5798b3655250940078691781571be42"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.782663 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" event={"ID":"332d7faf-04ea-49e5-8dfc-764444f702cb","Type":"ContainerDied","Data":"6b1f2646ba792b911c0d29e46a72b6face63fb159dff77e38cdc5d982fb8af83"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.782694 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jmlv2" Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.784092 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"8e8a9388-578d-4d41-a2cf-d2ad6e344288","Type":"ContainerStarted","Data":"c3753177a7ec40644ba71be484a5d282e419f9e7854cdb6133338516e599b83d"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.785950 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" event={"ID":"60517c3e-f430-453c-b6d4-3cdf124e100b","Type":"ContainerStarted","Data":"d648524567732f8f41f9735bf7cb84264f5634e30b2a138d94589820af03a16c"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.787692 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" event={"ID":"5cc4eeaa-88b7-4439-a39f-5e889845181e","Type":"ContainerStarted","Data":"a3853ddfc0b1612088f40e65300776863d775a61d36675b4d57f1f1541327b42"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.788692 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc","Type":"ContainerStarted","Data":"4064dc0440a4f2bfe7839354526eb059ccffe45f58c80ceb5401667fe760855c"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.790206 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7e7b2075-1317-4327-a8d9-4ebda5a68344","Type":"ContainerStarted","Data":"7b63b8911a6cc3f1de2643e90ff71a126d94fa08c9ca751441271681db8c061a"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.793614 4960 generic.go:334] "Generic (PLEG): container finished" podID="c107ba8b-ef9e-458f-a16f-f28f0625a85f" containerID="cde6d9062ff642ba91612f15ab42c084f387dc3d584433060fa1ef0ea0896e53" exitCode=0 Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.793731 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" event={"ID":"c107ba8b-ef9e-458f-a16f-f28f0625a85f","Type":"ContainerDied","Data":"cde6d9062ff642ba91612f15ab42c084f387dc3d584433060fa1ef0ea0896e53"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.793803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" event={"ID":"c107ba8b-ef9e-458f-a16f-f28f0625a85f","Type":"ContainerStarted","Data":"875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.795037 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" event={"ID":"2d2cf081-f13a-411b-bf24-5302a947dd9a","Type":"ContainerStarted","Data":"51d53e9a6ab0220c1f41abd50ab9bff8ce951970753b1134d143048b13d6909a"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.797292 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"39fbd706-99ef-4b5d-9d7a-e84ee337c443","Type":"ContainerStarted","Data":"e0be3b2ed1fe8c9679d308b4ed46cd1ff2f49c5f0601966a18385697ae0a6425"} Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.849091 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 20:00:08 crc kubenswrapper[4960]: I1201 20:00:08.856978 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jmlv2"] Dec 01 20:00:09 crc kubenswrapper[4960]: I1201 20:00:09.338641 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332d7faf-04ea-49e5-8dfc-764444f702cb" path="/var/lib/kubelet/pods/332d7faf-04ea-49e5-8dfc-764444f702cb/volumes" Dec 01 20:00:09 crc kubenswrapper[4960]: I1201 20:00:09.339480 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85bc78b4-f077-43b0-8a5a-5f56301a6af6" path="/var/lib/kubelet/pods/85bc78b4-f077-43b0-8a5a-5f56301a6af6/volumes" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.777861 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.819611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" event={"ID":"c107ba8b-ef9e-458f-a16f-f28f0625a85f","Type":"ContainerDied","Data":"875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc"} Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.819650 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="875573e095d79da57432bd3110367e3008e782bd518e95322b6b5baa066a53fc" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.819694 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.946561 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume\") pod \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.946797 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc6vl\" (UniqueName: \"kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl\") pod \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.946893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume\") pod \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\" (UID: \"c107ba8b-ef9e-458f-a16f-f28f0625a85f\") " Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.947554 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume" (OuterVolumeSpecName: "config-volume") pod "c107ba8b-ef9e-458f-a16f-f28f0625a85f" (UID: "c107ba8b-ef9e-458f-a16f-f28f0625a85f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.948249 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c107ba8b-ef9e-458f-a16f-f28f0625a85f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.954171 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c107ba8b-ef9e-458f-a16f-f28f0625a85f" (UID: "c107ba8b-ef9e-458f-a16f-f28f0625a85f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:00:10 crc kubenswrapper[4960]: I1201 20:00:10.954686 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl" (OuterVolumeSpecName: "kube-api-access-gc6vl") pod "c107ba8b-ef9e-458f-a16f-f28f0625a85f" (UID: "c107ba8b-ef9e-458f-a16f-f28f0625a85f"). InnerVolumeSpecName "kube-api-access-gc6vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:11 crc kubenswrapper[4960]: I1201 20:00:11.049768 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc6vl\" (UniqueName: \"kubernetes.io/projected/c107ba8b-ef9e-458f-a16f-f28f0625a85f-kube-api-access-gc6vl\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:11 crc kubenswrapper[4960]: I1201 20:00:11.049821 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c107ba8b-ef9e-458f-a16f-f28f0625a85f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:13 crc kubenswrapper[4960]: I1201 20:00:13.870526 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:00:13 crc kubenswrapper[4960]: I1201 20:00:13.871457 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:00:17 crc kubenswrapper[4960]: I1201 20:00:17.919204 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrf7z" event={"ID":"87e16436-45f5-4392-b524-b86f1cc5c0b2","Type":"ContainerStarted","Data":"a58c8bc705f60e8e057f2f6ee4a6c000be389dce35ea1ccb79d72f90c28cd462"} Dec 01 20:00:17 crc kubenswrapper[4960]: I1201 20:00:17.920151 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-nrf7z" Dec 01 20:00:17 crc kubenswrapper[4960]: I1201 20:00:17.940358 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nrf7z" podStartSLOduration=28.348351211 podStartE2EDuration="34.940334353s" podCreationTimestamp="2025-12-01 19:59:43 +0000 UTC" firstStartedPulling="2025-12-01 20:00:04.485571064 +0000 UTC m=+1239.773062773" lastFinishedPulling="2025-12-01 20:00:11.077554236 +0000 UTC m=+1246.365045915" observedRunningTime="2025-12-01 20:00:17.938830614 +0000 UTC m=+1253.226322303" watchObservedRunningTime="2025-12-01 20:00:17.940334353 +0000 UTC m=+1253.227826022" Dec 01 20:00:18 crc kubenswrapper[4960]: I1201 20:00:18.936293 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9n6l4" event={"ID":"9389e08d-0203-4f96-8278-fa1780a4201b","Type":"ContainerStarted","Data":"4eb52c2e1ddf09fd8bb2e53f2aff6b10dfeaf1226c01965df25ca5fb5c6048db"} Dec 01 20:00:19 crc kubenswrapper[4960]: I1201 20:00:19.943699 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a7eb168-4c2f-46c0-9292-de8e5c4383a5","Type":"ContainerStarted","Data":"912e0e4d62d15c9e862ab57b021bd64b2c31e51beaf443ac7af4c1d909eaf694"} Dec 01 20:00:19 crc kubenswrapper[4960]: I1201 20:00:19.945064 4960 generic.go:334] "Generic (PLEG): container finished" podID="9389e08d-0203-4f96-8278-fa1780a4201b" containerID="4eb52c2e1ddf09fd8bb2e53f2aff6b10dfeaf1226c01965df25ca5fb5c6048db" exitCode=0 Dec 01 20:00:19 crc kubenswrapper[4960]: I1201 20:00:19.945176 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9n6l4" event={"ID":"9389e08d-0203-4f96-8278-fa1780a4201b","Type":"ContainerDied","Data":"4eb52c2e1ddf09fd8bb2e53f2aff6b10dfeaf1226c01965df25ca5fb5c6048db"} Dec 01 20:00:19 crc kubenswrapper[4960]: I1201 20:00:19.947421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc","Type":"ContainerStarted","Data":"010b6be212fd65924e1990eacfd18195fa67fb579e28ddda60e309fdaf987e67"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.960948 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerStarted","Data":"277493da9f058effbde9f62ad5cd02fcd55a6dbc036097577f98bd4b1b9a3ed9"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.965124 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" event={"ID":"2d2cf081-f13a-411b-bf24-5302a947dd9a","Type":"ContainerStarted","Data":"03a2fb6b9bd92f8e57176635284a56dd38e87f5a2db09795e70b5e4c7b6de5d0"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.965633 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.968687 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"39fbd706-99ef-4b5d-9d7a-e84ee337c443","Type":"ContainerStarted","Data":"106fdede7b98228594fa27d27cfe9dcd6a06280eb16e78f26329aa44f9c08b06"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.978220 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-compactor-0" event={"ID":"8e8a9388-578d-4d41-a2cf-d2ad6e344288","Type":"ContainerStarted","Data":"dc26ae66f6607c13ac221d0e7bb943e6982be76c3f54a9b02d242bbb8063377c"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.979927 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.988181 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" event={"ID":"2e566703-0d27-4306-bd42-7a91da29c736","Type":"ContainerStarted","Data":"b92254e84b0c07dcf7ea3d50f96a231bbd14f828a0fbffabffdba2eacb1beb08"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.988288 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.998592 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6befffd7-b8c4-4c6c-900c-f6e1728f2a9e","Type":"ContainerStarted","Data":"4901be876732d70f7ca6b6dfebd15566cb21f8f6c9c74c1c295aae833dc44f0e"} Dec 01 20:00:20 crc kubenswrapper[4960]: I1201 20:00:20.999537 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 01 20:00:21 crc kubenswrapper[4960]: I1201 20:00:21.034270 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-compactor-0" podStartSLOduration=22.96026686 podStartE2EDuration="33.034252714s" podCreationTimestamp="2025-12-01 19:59:48 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.768697115 +0000 UTC m=+1243.056188784" lastFinishedPulling="2025-12-01 20:00:17.842682969 +0000 UTC m=+1253.130174638" observedRunningTime="2025-12-01 20:00:21.03103359 +0000 UTC m=+1256.318525259" watchObservedRunningTime="2025-12-01 20:00:21.034252714 +0000 UTC m=+1256.321744383" Dec 01 20:00:21 crc kubenswrapper[4960]: I1201 20:00:21.061370 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" podStartSLOduration=22.922803123 podStartE2EDuration="33.06134951s" podCreationTimestamp="2025-12-01 19:59:48 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.726619878 +0000 UTC m=+1243.014111547" lastFinishedPulling="2025-12-01 20:00:17.865166265 +0000 UTC m=+1253.152657934" observedRunningTime="2025-12-01 20:00:21.052986422 +0000 UTC m=+1256.340478091" watchObservedRunningTime="2025-12-01 20:00:21.06134951 +0000 UTC m=+1256.348841179" Dec 01 20:00:21 crc kubenswrapper[4960]: I1201 20:00:21.076998 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" podStartSLOduration=22.134380176 podStartE2EDuration="32.076974677s" podCreationTimestamp="2025-12-01 19:59:49 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.754216027 +0000 UTC m=+1243.041707706" lastFinishedPulling="2025-12-01 20:00:17.696810528 +0000 UTC m=+1252.984302207" observedRunningTime="2025-12-01 20:00:21.072313916 +0000 UTC m=+1256.359805595" watchObservedRunningTime="2025-12-01 20:00:21.076974677 +0000 UTC m=+1256.364466346" Dec 01 20:00:21 crc kubenswrapper[4960]: I1201 20:00:21.099835 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=4.863254044 podStartE2EDuration="44.099812792s" podCreationTimestamp="2025-12-01 19:59:37 +0000 UTC" firstStartedPulling="2025-12-01 19:59:38.463467354 +0000 UTC m=+1213.750959023" lastFinishedPulling="2025-12-01 20:00:17.700026102 +0000 UTC m=+1252.987517771" observedRunningTime="2025-12-01 20:00:21.093716593 +0000 UTC m=+1256.381208272" watchObservedRunningTime="2025-12-01 20:00:21.099812792 +0000 UTC m=+1256.387304461" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.011707 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2","Type":"ContainerStarted","Data":"47a726ec89f0e917db67dfa3996ecc85eb687ed4c2cac0f3aa9671bd1019461f"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.014028 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02","Type":"ContainerStarted","Data":"daadb39326e4447e5d9be3faa404a30ae35d3269d9bf451d2f4d24d06afdb628"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.018380 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-ingester-0" event={"ID":"969241f0-8883-4570-94d4-fdb20ee094cd","Type":"ContainerStarted","Data":"5e0d4797768f2bdf80004fa1c81da47d94798402a30a49559395a8303575180d"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.019022 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.021138 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" event={"ID":"a8b0ac50-f375-427e-b103-0bfd272201ea","Type":"ContainerStarted","Data":"be4b322953343c88596547480001f655e225b601e3f27766b4d779f0858bbd55"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.021761 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.024205 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" event={"ID":"60517c3e-f430-453c-b6d4-3cdf124e100b","Type":"ContainerStarted","Data":"72ff8b4c0fb54a92f9c8909ce36423c9d639faa3ae1d5555fae9d2aa9acdcf28"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.024792 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.027926 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-index-gateway-0" event={"ID":"eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38","Type":"ContainerStarted","Data":"c437d2326c1e2aaa28d4282c9fc175690bfb018c87bf1ac63e8774f86ae27a9a"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.028167 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.031105 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" event={"ID":"5cc4eeaa-88b7-4439-a39f-5e889845181e","Type":"ContainerStarted","Data":"962ed9384821964b7cb762f035d4a2e5f4c6f6d779d4afb2ece799286f747118"} Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.047099 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.123659 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" podStartSLOduration=23.741232267 podStartE2EDuration="34.123639197s" podCreationTimestamp="2025-12-01 19:59:48 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.740699315 +0000 UTC m=+1243.028190984" lastFinishedPulling="2025-12-01 20:00:18.123106245 +0000 UTC m=+1253.410597914" observedRunningTime="2025-12-01 20:00:22.089686003 +0000 UTC m=+1257.377177672" watchObservedRunningTime="2025-12-01 20:00:22.123639197 +0000 UTC m=+1257.411130866" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.127547 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-index-gateway-0" podStartSLOduration=22.104223241 podStartE2EDuration="33.127531909s" podCreationTimestamp="2025-12-01 19:59:49 +0000 UTC" firstStartedPulling="2025-12-01 20:00:06.802827089 +0000 UTC m=+1242.090318758" lastFinishedPulling="2025-12-01 20:00:17.826135757 +0000 UTC m=+1253.113627426" observedRunningTime="2025-12-01 20:00:22.109872918 +0000 UTC m=+1257.397364607" watchObservedRunningTime="2025-12-01 20:00:22.127531909 +0000 UTC m=+1257.415023578" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.147254 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-xw4bs" podStartSLOduration=23.486546036 podStartE2EDuration="33.147231262s" podCreationTimestamp="2025-12-01 19:59:49 +0000 UTC" firstStartedPulling="2025-12-01 20:00:08.036196284 +0000 UTC m=+1243.323687953" lastFinishedPulling="2025-12-01 20:00:17.69688151 +0000 UTC m=+1252.984373179" observedRunningTime="2025-12-01 20:00:22.133910215 +0000 UTC m=+1257.421401894" watchObservedRunningTime="2025-12-01 20:00:22.147231262 +0000 UTC m=+1257.434722931" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.167378 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" podStartSLOduration=22.476292275 podStartE2EDuration="33.167356437s" podCreationTimestamp="2025-12-01 19:59:49 +0000 UTC" firstStartedPulling="2025-12-01 20:00:08.041373059 +0000 UTC m=+1243.328864728" lastFinishedPulling="2025-12-01 20:00:18.732437181 +0000 UTC m=+1254.019928890" observedRunningTime="2025-12-01 20:00:22.153900116 +0000 UTC m=+1257.441391805" watchObservedRunningTime="2025-12-01 20:00:22.167356437 +0000 UTC m=+1257.454848106" Dec 01 20:00:22 crc kubenswrapper[4960]: I1201 20:00:22.186890 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-lokistack-ingester-0" podStartSLOduration=24.215291029 podStartE2EDuration="34.186872575s" podCreationTimestamp="2025-12-01 19:59:48 +0000 UTC" firstStartedPulling="2025-12-01 20:00:08.023850383 +0000 UTC m=+1243.311342052" lastFinishedPulling="2025-12-01 20:00:17.995431889 +0000 UTC m=+1253.282923598" observedRunningTime="2025-12-01 20:00:22.173659421 +0000 UTC m=+1257.461151090" watchObservedRunningTime="2025-12-01 20:00:22.186872575 +0000 UTC m=+1257.474364244" Dec 01 20:00:23 crc kubenswrapper[4960]: I1201 20:00:23.041406 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerStarted","Data":"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577"} Dec 01 20:00:23 crc kubenswrapper[4960]: I1201 20:00:23.044518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9n6l4" event={"ID":"9389e08d-0203-4f96-8278-fa1780a4201b","Type":"ContainerStarted","Data":"005e888a359db76051af823c1dd82be1951d5415e131300156cc724435ec1877"} Dec 01 20:00:23 crc kubenswrapper[4960]: I1201 20:00:23.047713 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerStarted","Data":"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5"} Dec 01 20:00:23 crc kubenswrapper[4960]: I1201 20:00:23.048709 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 20:00:23 crc kubenswrapper[4960]: I1201 20:00:23.069176 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-gateway-76cc998948-mpddm" Dec 01 20:00:27 crc kubenswrapper[4960]: I1201 20:00:27.084698 4960 generic.go:334] "Generic (PLEG): container finished" podID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerID="277493da9f058effbde9f62ad5cd02fcd55a6dbc036097577f98bd4b1b9a3ed9" exitCode=0 Dec 01 20:00:27 crc kubenswrapper[4960]: I1201 20:00:27.084794 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerDied","Data":"277493da9f058effbde9f62ad5cd02fcd55a6dbc036097577f98bd4b1b9a3ed9"} Dec 01 20:00:27 crc kubenswrapper[4960]: I1201 20:00:27.814525 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.103874 4960 generic.go:334] "Generic (PLEG): container finished" podID="39fbd706-99ef-4b5d-9d7a-e84ee337c443" containerID="106fdede7b98228594fa27d27cfe9dcd6a06280eb16e78f26329aa44f9c08b06" exitCode=0 Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.103989 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"39fbd706-99ef-4b5d-9d7a-e84ee337c443","Type":"ContainerDied","Data":"106fdede7b98228594fa27d27cfe9dcd6a06280eb16e78f26329aa44f9c08b06"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.110631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-9n6l4" event={"ID":"9389e08d-0203-4f96-8278-fa1780a4201b","Type":"ContainerStarted","Data":"95d557d31947d0c9c221e2bdd1d57e8b2844e70472197ac0ff8b1f6e942f39a6"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.111592 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.111627 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.116869 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4a7eb168-4c2f-46c0-9292-de8e5c4383a5","Type":"ContainerStarted","Data":"c34cb456d1d80fc31b7c9ce75a8fdcbdc2ee2d2b48cf005ea135712fe2cd9c3e"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.119669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1c6f94f2-a256-4e77-b6d1-39706b7c6ddc","Type":"ContainerStarted","Data":"de1925a9ae18de3367bbb07df924e84eef465a0df325a789fe5020896a343d97"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.121666 4960 generic.go:334] "Generic (PLEG): container finished" podID="f9b50b6f-1af6-4283-8994-1cdc2b9e6c02" containerID="daadb39326e4447e5d9be3faa404a30ae35d3269d9bf451d2f4d24d06afdb628" exitCode=0 Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.121735 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02","Type":"ContainerDied","Data":"daadb39326e4447e5d9be3faa404a30ae35d3269d9bf451d2f4d24d06afdb628"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.123334 4960 generic.go:334] "Generic (PLEG): container finished" podID="b484313f-1265-4ac3-8d73-5dbe81926118" containerID="357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd" exitCode=0 Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.123400 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" event={"ID":"b484313f-1265-4ac3-8d73-5dbe81926118","Type":"ContainerDied","Data":"357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.126636 4960 generic.go:334] "Generic (PLEG): container finished" podID="795f6715-b4c6-410f-868c-eeb7139913d1" containerID="6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7" exitCode=0 Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.126707 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" event={"ID":"795f6715-b4c6-410f-868c-eeb7139913d1","Type":"ContainerDied","Data":"6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.131682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7e7b2075-1317-4327-a8d9-4ebda5a68344","Type":"ContainerStarted","Data":"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.132444 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.134931 4960 generic.go:334] "Generic (PLEG): container finished" podID="c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2" containerID="47a726ec89f0e917db67dfa3996ecc85eb687ed4c2cac0f3aa9671bd1019461f" exitCode=0 Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.134984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2","Type":"ContainerDied","Data":"47a726ec89f0e917db67dfa3996ecc85eb687ed4c2cac0f3aa9671bd1019461f"} Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.164958 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-9n6l4" podStartSLOduration=41.664793585 podStartE2EDuration="45.164934681s" podCreationTimestamp="2025-12-01 19:59:43 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.611031106 +0000 UTC m=+1242.898522775" lastFinishedPulling="2025-12-01 20:00:11.111172212 +0000 UTC m=+1246.398663871" observedRunningTime="2025-12-01 20:00:28.159988563 +0000 UTC m=+1263.447480232" watchObservedRunningTime="2025-12-01 20:00:28.164934681 +0000 UTC m=+1263.452426350" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.185669 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.278910982 podStartE2EDuration="42.185653181s" podCreationTimestamp="2025-12-01 19:59:46 +0000 UTC" firstStartedPulling="2025-12-01 20:00:08.089596415 +0000 UTC m=+1243.377088084" lastFinishedPulling="2025-12-01 20:00:26.996338614 +0000 UTC m=+1262.283830283" observedRunningTime="2025-12-01 20:00:28.1813401 +0000 UTC m=+1263.468831769" watchObservedRunningTime="2025-12-01 20:00:28.185653181 +0000 UTC m=+1263.473144850" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.197730 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=30.349523702 podStartE2EDuration="49.197707666s" podCreationTimestamp="2025-12-01 19:59:39 +0000 UTC" firstStartedPulling="2025-12-01 20:00:08.04182918 +0000 UTC m=+1243.329320849" lastFinishedPulling="2025-12-01 20:00:26.890013144 +0000 UTC m=+1262.177504813" observedRunningTime="2025-12-01 20:00:28.195492178 +0000 UTC m=+1263.482983847" watchObservedRunningTime="2025-12-01 20:00:28.197707666 +0000 UTC m=+1263.485199345" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.238147 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=27.107883576 podStartE2EDuration="46.238125698s" podCreationTimestamp="2025-12-01 19:59:42 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.879714857 +0000 UTC m=+1243.167206526" lastFinishedPulling="2025-12-01 20:00:27.009956979 +0000 UTC m=+1262.297448648" observedRunningTime="2025-12-01 20:00:28.233817387 +0000 UTC m=+1263.521309056" watchObservedRunningTime="2025-12-01 20:00:28.238125698 +0000 UTC m=+1263.525617367" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.391582 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.851468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.851525 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 01 20:00:28 crc kubenswrapper[4960]: I1201 20:00:28.936950 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.148747 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" event={"ID":"b484313f-1265-4ac3-8d73-5dbe81926118","Type":"ContainerStarted","Data":"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38"} Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.148976 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.151372 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" event={"ID":"795f6715-b4c6-410f-868c-eeb7139913d1","Type":"ContainerStarted","Data":"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3"} Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.151587 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.153863 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2","Type":"ContainerStarted","Data":"89c3b829031427336b15ec1d0d4fd3bb7d7ce483c188eec5347ac98623f4eeea"} Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.156472 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f9b50b6f-1af6-4283-8994-1cdc2b9e6c02","Type":"ContainerStarted","Data":"998d780bd8f7e7e76aab18c15d5bbb6dcaf159e78b51ecde4cd7cc615d786593"} Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.173516 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" podStartSLOduration=4.194029998 podStartE2EDuration="57.173495819s" podCreationTimestamp="2025-12-01 19:59:32 +0000 UTC" firstStartedPulling="2025-12-01 19:59:34.007333784 +0000 UTC m=+1209.294825453" lastFinishedPulling="2025-12-01 20:00:26.986799605 +0000 UTC m=+1262.274291274" observedRunningTime="2025-12-01 20:00:29.168403657 +0000 UTC m=+1264.455895346" watchObservedRunningTime="2025-12-01 20:00:29.173495819 +0000 UTC m=+1264.460987488" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.190806 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" podStartSLOduration=4.176164271 podStartE2EDuration="57.19078527s" podCreationTimestamp="2025-12-01 19:59:32 +0000 UTC" firstStartedPulling="2025-12-01 19:59:33.994632271 +0000 UTC m=+1209.282123940" lastFinishedPulling="2025-12-01 20:00:27.00925327 +0000 UTC m=+1262.296744939" observedRunningTime="2025-12-01 20:00:29.188033838 +0000 UTC m=+1264.475525507" watchObservedRunningTime="2025-12-01 20:00:29.19078527 +0000 UTC m=+1264.478276939" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.213665 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.219136 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=13.821672255 podStartE2EDuration="54.219097328s" podCreationTimestamp="2025-12-01 19:59:35 +0000 UTC" firstStartedPulling="2025-12-01 19:59:38.314226176 +0000 UTC m=+1213.601717845" lastFinishedPulling="2025-12-01 20:00:18.711651239 +0000 UTC m=+1253.999142918" observedRunningTime="2025-12-01 20:00:29.21036632 +0000 UTC m=+1264.497857979" watchObservedRunningTime="2025-12-01 20:00:29.219097328 +0000 UTC m=+1264.506588997" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.241301 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.798383794 podStartE2EDuration="55.241276906s" podCreationTimestamp="2025-12-01 19:59:34 +0000 UTC" firstStartedPulling="2025-12-01 19:59:36.54309271 +0000 UTC m=+1211.830584379" lastFinishedPulling="2025-12-01 20:00:17.985985822 +0000 UTC m=+1253.273477491" observedRunningTime="2025-12-01 20:00:29.232738803 +0000 UTC m=+1264.520230482" watchObservedRunningTime="2025-12-01 20:00:29.241276906 +0000 UTC m=+1264.528768575" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.576752 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.637394 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:29 crc kubenswrapper[4960]: E1201 20:00:29.637842 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c107ba8b-ef9e-458f-a16f-f28f0625a85f" containerName="collect-profiles" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.637859 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c107ba8b-ef9e-458f-a16f-f28f0625a85f" containerName="collect-profiles" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.638027 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c107ba8b-ef9e-458f-a16f-f28f0625a85f" containerName="collect-profiles" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.639093 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.645228 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.649597 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-trsdq"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.659084 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.659823 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-trsdq"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.661233 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.688195 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.707863 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728179 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728208 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728222 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hnp4\" (UniqueName: \"kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728245 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdjcp\" (UniqueName: \"kubernetes.io/projected/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-kube-api-access-pdjcp\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728273 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728320 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-config\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728347 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-combined-ca-bundle\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728381 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovs-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.728400 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovn-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.757747 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.759798 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.804768 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830027 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-config\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830080 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-combined-ca-bundle\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830350 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovs-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830379 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovn-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830450 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830478 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830499 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830521 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830537 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hnp4\" (UniqueName: \"kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdjcp\" (UniqueName: \"kubernetes.io/projected/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-kube-api-access-pdjcp\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830589 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830623 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7fkt\" (UniqueName: \"kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.830644 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.831573 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.832031 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.832353 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.832604 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovn-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.832624 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-ovs-rundir\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.832726 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-config\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.843192 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-combined-ca-bundle\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.855640 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.858922 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdjcp\" (UniqueName: \"kubernetes.io/projected/5c1ff07c-096b-4952-a0cb-32fdcfaf154a-kube-api-access-pdjcp\") pod \"ovn-controller-metrics-trsdq\" (UID: \"5c1ff07c-096b-4952-a0cb-32fdcfaf154a\") " pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.860722 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hnp4\" (UniqueName: \"kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4\") pod \"dnsmasq-dns-7fd796d7df-t7hrw\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.932639 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7fkt\" (UniqueName: \"kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.932704 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.932852 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.932898 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.933999 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.934174 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.936795 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:29 crc kubenswrapper[4960]: I1201 20:00:29.957012 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7fkt\" (UniqueName: \"kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt\") pod \"dnsmasq-dns-74f6f696b9-8j49h\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.018571 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.043880 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-trsdq" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.058446 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.061962 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.099190 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.101199 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.111210 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.113250 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.238409 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s58r\" (UniqueName: \"kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.238779 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.238891 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.238928 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.238962 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.346200 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.346892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.346940 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.347076 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s58r\" (UniqueName: \"kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.347102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.348391 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.349772 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.350527 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.350823 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.374665 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s58r\" (UniqueName: \"kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r\") pod \"dnsmasq-dns-698758b865-p7pmg\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.390887 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.446309 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.467108 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.575362 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.812424 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.825205 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.832270 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.832390 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-g6gtj" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.832544 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.833605 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.840981 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-trsdq"] Dec 01 20:00:30 crc kubenswrapper[4960]: W1201 20:00:30.851830 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc2dbc80_e228_4c1e_a06a_ae8ab6d71910.slice/crio-9315c0362ffcb2ca2b23fbce8b116643b3b77b31ea28a9f229d99f45ce88478b WatchSource:0}: Error finding container 9315c0362ffcb2ca2b23fbce8b116643b3b77b31ea28a9f229d99f45ce88478b: Status 404 returned error can't find the container with id 9315c0362ffcb2ca2b23fbce8b116643b3b77b31ea28a9f229d99f45ce88478b Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.865073 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.865325 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrphl\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-kube-api-access-zrphl\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.865361 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-lock\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.865392 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.865445 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-cache\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.872824 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.892303 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.970892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrphl\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-kube-api-access-zrphl\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.970973 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-lock\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.971025 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.971097 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-cache\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.971385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: E1201 20:00:30.971676 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:30 crc kubenswrapper[4960]: E1201 20:00:30.971699 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:30 crc kubenswrapper[4960]: E1201 20:00:30.971753 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:00:31.471732202 +0000 UTC m=+1266.759223871 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.972129 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-lock\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.973816 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/725698af-f23c-4262-96a4-2bfff12018fa-cache\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.981345 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.981388 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/596061f7c665cf6f04527023c12d65ccdec52475e6ca4ace9c96c9e4ec425812/globalmount\"" pod="openstack/swift-storage-0" Dec 01 20:00:30 crc kubenswrapper[4960]: I1201 20:00:30.996976 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrphl\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-kube-api-access-zrphl\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.046614 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-681492e7-a1ab-4d03-baa2-ad16e10b0a06\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.154547 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.224814 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-trsdq" event={"ID":"5c1ff07c-096b-4952-a0cb-32fdcfaf154a","Type":"ContainerStarted","Data":"c811ee0b979a458ce1d09b0b0276e0c09f78dbd66c57c39b6c81bc7f22beeed9"} Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.238880 4960 generic.go:334] "Generic (PLEG): container finished" podID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerID="fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed" exitCode=0 Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.238975 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" event={"ID":"001b0c13-5e8e-4458-839a-83059a9bd06f","Type":"ContainerDied","Data":"fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed"} Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.239008 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" event={"ID":"001b0c13-5e8e-4458-839a-83059a9bd06f","Type":"ContainerStarted","Data":"2b8a05acb81b100790909277aaa3dd0dcfe368cdd3d1a9233d45e3e3c2d4bb7a"} Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.242831 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" event={"ID":"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910","Type":"ContainerStarted","Data":"9315c0362ffcb2ca2b23fbce8b116643b3b77b31ea28a9f229d99f45ce88478b"} Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.244495 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p7pmg" event={"ID":"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d","Type":"ContainerStarted","Data":"a52d94d6ccd24ac7f0b3db47a152423a89224b66104078690d5c8c2fa2195a8b"} Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.245213 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="dnsmasq-dns" containerID="cri-o://d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38" gracePeriod=10 Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.245373 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="dnsmasq-dns" containerID="cri-o://773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3" gracePeriod=10 Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.360670 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-gmqjb"] Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.362052 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.364730 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.368671 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gmqjb"] Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.373908 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.381078 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.385029 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484006 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484070 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484103 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rtvx\" (UniqueName: \"kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484151 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484228 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484298 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.484421 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: E1201 20:00:31.486769 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:31 crc kubenswrapper[4960]: E1201 20:00:31.486811 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:31 crc kubenswrapper[4960]: E1201 20:00:31.486877 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:00:32.486851194 +0000 UTC m=+1267.774342913 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.587544 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588177 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rtvx\" (UniqueName: \"kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588212 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588273 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588327 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588392 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588423 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.588444 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.589481 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.591603 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.592182 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.592503 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.597274 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-np65t" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.597488 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.597597 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.598778 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.603966 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.616461 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.620053 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.620584 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.647032 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rtvx\" (UniqueName: \"kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx\") pod \"swift-ring-rebalance-gmqjb\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694153 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-config\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694476 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694531 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694561 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07787692-57cc-473c-ab9b-276004fb9220-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694618 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdc4j\" (UniqueName: \"kubernetes.io/projected/07787692-57cc-473c-ab9b-276004fb9220-kube-api-access-hdc4j\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.694640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-scripts\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.722035 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.763755 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.800168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-config\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.800708 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.800873 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.800957 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.800988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07787692-57cc-473c-ab9b-276004fb9220-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.801015 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-config\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.802729 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07787692-57cc-473c-ab9b-276004fb9220-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.802826 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdc4j\" (UniqueName: \"kubernetes.io/projected/07787692-57cc-473c-ab9b-276004fb9220-kube-api-access-hdc4j\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.802883 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-scripts\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.808105 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.808428 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.808547 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07787692-57cc-473c-ab9b-276004fb9220-scripts\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.811156 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/07787692-57cc-473c-ab9b-276004fb9220-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.831741 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdc4j\" (UniqueName: \"kubernetes.io/projected/07787692-57cc-473c-ab9b-276004fb9220-kube-api-access-hdc4j\") pod \"ovn-northd-0\" (UID: \"07787692-57cc-473c-ab9b-276004fb9220\") " pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.911206 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68pht\" (UniqueName: \"kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht\") pod \"795f6715-b4c6-410f-868c-eeb7139913d1\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.911312 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc\") pod \"795f6715-b4c6-410f-868c-eeb7139913d1\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.911362 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config\") pod \"795f6715-b4c6-410f-868c-eeb7139913d1\" (UID: \"795f6715-b4c6-410f-868c-eeb7139913d1\") " Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.916773 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht" (OuterVolumeSpecName: "kube-api-access-68pht") pod "795f6715-b4c6-410f-868c-eeb7139913d1" (UID: "795f6715-b4c6-410f-868c-eeb7139913d1"). InnerVolumeSpecName "kube-api-access-68pht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.941842 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 01 20:00:31 crc kubenswrapper[4960]: I1201 20:00:31.980459 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config" (OuterVolumeSpecName: "config") pod "795f6715-b4c6-410f-868c-eeb7139913d1" (UID: "795f6715-b4c6-410f-868c-eeb7139913d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.013432 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68pht\" (UniqueName: \"kubernetes.io/projected/795f6715-b4c6-410f-868c-eeb7139913d1-kube-api-access-68pht\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.013457 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.023774 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "795f6715-b4c6-410f-868c-eeb7139913d1" (UID: "795f6715-b4c6-410f-868c-eeb7139913d1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.104491 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.114865 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/795f6715-b4c6-410f-868c-eeb7139913d1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.216394 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc\") pod \"b484313f-1265-4ac3-8d73-5dbe81926118\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.216605 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config\") pod \"b484313f-1265-4ac3-8d73-5dbe81926118\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.216662 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm5m6\" (UniqueName: \"kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6\") pod \"b484313f-1265-4ac3-8d73-5dbe81926118\" (UID: \"b484313f-1265-4ac3-8d73-5dbe81926118\") " Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.224295 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6" (OuterVolumeSpecName: "kube-api-access-mm5m6") pod "b484313f-1265-4ac3-8d73-5dbe81926118" (UID: "b484313f-1265-4ac3-8d73-5dbe81926118"). InnerVolumeSpecName "kube-api-access-mm5m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.284730 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" event={"ID":"001b0c13-5e8e-4458-839a-83059a9bd06f","Type":"ContainerStarted","Data":"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.285330 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.285976 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config" (OuterVolumeSpecName: "config") pod "b484313f-1265-4ac3-8d73-5dbe81926118" (UID: "b484313f-1265-4ac3-8d73-5dbe81926118"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.294518 4960 generic.go:334] "Generic (PLEG): container finished" podID="b484313f-1265-4ac3-8d73-5dbe81926118" containerID="d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38" exitCode=0 Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.294598 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" event={"ID":"b484313f-1265-4ac3-8d73-5dbe81926118","Type":"ContainerDied","Data":"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.294634 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" event={"ID":"b484313f-1265-4ac3-8d73-5dbe81926118","Type":"ContainerDied","Data":"3d4fd8b6f9fb2dcbfc1030328e31ac117aa6bb4178cde8d112baa68930dd4100"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.294654 4960 scope.go:117] "RemoveContainer" containerID="d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.294799 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-qmxlv" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.297885 4960 generic.go:334] "Generic (PLEG): container finished" podID="dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" containerID="49cb918750abfe29b83b2438c9ec0942958e56e4f5fcfa617cf1ae7d773150aa" exitCode=0 Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.297942 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" event={"ID":"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910","Type":"ContainerDied","Data":"49cb918750abfe29b83b2438c9ec0942958e56e4f5fcfa617cf1ae7d773150aa"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.310854 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b484313f-1265-4ac3-8d73-5dbe81926118" (UID: "b484313f-1265-4ac3-8d73-5dbe81926118"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.311333 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" podStartSLOduration=3.311312804 podStartE2EDuration="3.311312804s" podCreationTimestamp="2025-12-01 20:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:00:32.308349207 +0000 UTC m=+1267.595840876" watchObservedRunningTime="2025-12-01 20:00:32.311312804 +0000 UTC m=+1267.598804473" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.319666 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm5m6\" (UniqueName: \"kubernetes.io/projected/b484313f-1265-4ac3-8d73-5dbe81926118-kube-api-access-mm5m6\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.319731 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.319741 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b484313f-1265-4ac3-8d73-5dbe81926118-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.326322 4960 generic.go:334] "Generic (PLEG): container finished" podID="795f6715-b4c6-410f-868c-eeb7139913d1" containerID="773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3" exitCode=0 Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.326722 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" event={"ID":"795f6715-b4c6-410f-868c-eeb7139913d1","Type":"ContainerDied","Data":"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.326752 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" event={"ID":"795f6715-b4c6-410f-868c-eeb7139913d1","Type":"ContainerDied","Data":"388745035a30e18a1cd72a2b7efad43e05ecc28a80d8ee04538645765ae77c7d"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.326828 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-zztmq" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.340212 4960 generic.go:334] "Generic (PLEG): container finished" podID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerID="1d403eec1c842f6b6acf20822bb58f7e2e682111b6d21a117d9d496e6dde017a" exitCode=0 Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.340949 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p7pmg" event={"ID":"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d","Type":"ContainerDied","Data":"1d403eec1c842f6b6acf20822bb58f7e2e682111b6d21a117d9d496e6dde017a"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.351060 4960 scope.go:117] "RemoveContainer" containerID="357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.353912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-trsdq" event={"ID":"5c1ff07c-096b-4952-a0cb-32fdcfaf154a","Type":"ContainerStarted","Data":"e70e42ca39d98ce06e6ff6862d6c23b1ae38642ecf9000f7695dd32babc3b555"} Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.394185 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.397826 4960 scope.go:117] "RemoveContainer" containerID="d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38" Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.398300 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38\": container with ID starting with d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38 not found: ID does not exist" containerID="d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.398328 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38"} err="failed to get container status \"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38\": rpc error: code = NotFound desc = could not find container \"d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38\": container with ID starting with d02060faa68e4a5a908eccc31e2dde1d07780425f9b10ea69962d90555626b38 not found: ID does not exist" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.398351 4960 scope.go:117] "RemoveContainer" containerID="357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd" Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.401900 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd\": container with ID starting with 357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd not found: ID does not exist" containerID="357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.401929 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd"} err="failed to get container status \"357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd\": rpc error: code = NotFound desc = could not find container \"357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd\": container with ID starting with 357f17bfd15f11495f575ec6348e3ca1702955c62bf3036fa8298c6e067e7cdd not found: ID does not exist" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.401950 4960 scope.go:117] "RemoveContainer" containerID="773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.405994 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-zztmq"] Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.442515 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gmqjb"] Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.447334 4960 scope.go:117] "RemoveContainer" containerID="6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.494410 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-trsdq" podStartSLOduration=3.492482215 podStartE2EDuration="3.492482215s" podCreationTimestamp="2025-12-01 20:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:00:32.415512339 +0000 UTC m=+1267.703004028" watchObservedRunningTime="2025-12-01 20:00:32.492482215 +0000 UTC m=+1267.779973884" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.523296 4960 scope.go:117] "RemoveContainer" containerID="773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3" Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.532012 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3\": container with ID starting with 773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3 not found: ID does not exist" containerID="773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.532092 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3"} err="failed to get container status \"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3\": rpc error: code = NotFound desc = could not find container \"773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3\": container with ID starting with 773122349a3f55b8424902f3f0b69f984a1abffb32eee0d31a631ce69c5e84c3 not found: ID does not exist" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.532170 4960 scope.go:117] "RemoveContainer" containerID="6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7" Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.532587 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7\": container with ID starting with 6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7 not found: ID does not exist" containerID="6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.532618 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7"} err="failed to get container status \"6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7\": rpc error: code = NotFound desc = could not find container \"6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7\": container with ID starting with 6f57092dc037270116ab9e8d49cbb04017b4c7ccab8597251edd251359e4d1d7 not found: ID does not exist" Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.542701 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.551594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.551945 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.551967 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:32 crc kubenswrapper[4960]: E1201 20:00:32.552003 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:00:34.551987595 +0000 UTC m=+1269.839479264 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.938665 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 20:00:32 crc kubenswrapper[4960]: I1201 20:00:32.951229 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-qmxlv"] Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.340704 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" path="/var/lib/kubelet/pods/795f6715-b4c6-410f-868c-eeb7139913d1/volumes" Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.341331 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" path="/var/lib/kubelet/pods/b484313f-1265-4ac3-8d73-5dbe81926118/volumes" Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.368612 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p7pmg" event={"ID":"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d","Type":"ContainerStarted","Data":"4abbc71a786a06dda888cbd461b9fa2d992debe605a22a578320c0d407cb1ddb"} Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.368775 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.370682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmqjb" event={"ID":"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe","Type":"ContainerStarted","Data":"98dfe4797cf0534f9d66b999e38dbd5c8e851436226a4ce01eb4b156fe6ee504"} Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.372383 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"07787692-57cc-473c-ab9b-276004fb9220","Type":"ContainerStarted","Data":"2ef9a79ae351ead94ce72ff3736137df63aea9e5162bc559641eba5e3c2da852"} Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.911988 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:33 crc kubenswrapper[4960]: I1201 20:00:33.938632 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-p7pmg" podStartSLOduration=3.938614363 podStartE2EDuration="3.938614363s" podCreationTimestamp="2025-12-01 20:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:00:33.392738861 +0000 UTC m=+1268.680230520" watchObservedRunningTime="2025-12-01 20:00:33.938614363 +0000 UTC m=+1269.226106032" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.097579 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config\") pod \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.097673 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc\") pod \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.097702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb\") pod \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.098697 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7fkt\" (UniqueName: \"kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt\") pod \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\" (UID: \"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910\") " Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.112058 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt" (OuterVolumeSpecName: "kube-api-access-z7fkt") pod "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" (UID: "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910"). InnerVolumeSpecName "kube-api-access-z7fkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.123305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" (UID: "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.135825 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config" (OuterVolumeSpecName: "config") pod "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" (UID: "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.151277 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" (UID: "dc2dbc80-e228-4c1e-a06a-ae8ab6d71910"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.201487 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.201529 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.201540 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.201554 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7fkt\" (UniqueName: \"kubernetes.io/projected/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910-kube-api-access-z7fkt\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.391266 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"39fbd706-99ef-4b5d-9d7a-e84ee337c443","Type":"ContainerStarted","Data":"7e446284b471f5cd858a3d92acf4db8a420398061e3fc0a9d6eef1a4d1887156"} Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.393948 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" event={"ID":"dc2dbc80-e228-4c1e-a06a-ae8ab6d71910","Type":"ContainerDied","Data":"9315c0362ffcb2ca2b23fbce8b116643b3b77b31ea28a9f229d99f45ce88478b"} Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.394012 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-8j49h" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.394025 4960 scope.go:117] "RemoveContainer" containerID="49cb918750abfe29b83b2438c9ec0942958e56e4f5fcfa617cf1ae7d773150aa" Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.470144 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.484898 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-8j49h"] Dec 01 20:00:34 crc kubenswrapper[4960]: I1201 20:00:34.608127 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:34 crc kubenswrapper[4960]: E1201 20:00:34.608524 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:34 crc kubenswrapper[4960]: E1201 20:00:34.608539 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:34 crc kubenswrapper[4960]: E1201 20:00:34.608587 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:00:38.608571119 +0000 UTC m=+1273.896062788 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:35 crc kubenswrapper[4960]: I1201 20:00:35.340886 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" path="/var/lib/kubelet/pods/dc2dbc80-e228-4c1e-a06a-ae8ab6d71910/volumes" Dec 01 20:00:36 crc kubenswrapper[4960]: I1201 20:00:35.999995 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 01 20:00:36 crc kubenswrapper[4960]: I1201 20:00:36.000178 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 01 20:00:36 crc kubenswrapper[4960]: I1201 20:00:36.256573 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 01 20:00:36 crc kubenswrapper[4960]: I1201 20:00:36.511064 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.399175 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2bwq2"] Dec 01 20:00:37 crc kubenswrapper[4960]: E1201 20:00:37.401569 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.401691 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: E1201 20:00:37.401787 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.401866 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: E1201 20:00:37.401949 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.402023 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: E1201 20:00:37.402152 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.402237 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: E1201 20:00:37.402319 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.402402 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.403138 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b484313f-1265-4ac3-8d73-5dbe81926118" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.403241 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2dbc80-e228-4c1e-a06a-ae8ab6d71910" containerName="init" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.403344 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="795f6715-b4c6-410f-868c-eeb7139913d1" containerName="dnsmasq-dns" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.404309 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.409053 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2bwq2"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.525677 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fbe8-account-create-update-2ngk5"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.527424 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.529262 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.533685 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fbe8-account-create-update-2ngk5"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.562386 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26dt2\" (UniqueName: \"kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.562543 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.590234 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.590282 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.664210 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.664237 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26dt2\" (UniqueName: \"kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.664353 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.664387 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.664464 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzv4h\" (UniqueName: \"kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.665660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.702970 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26dt2\" (UniqueName: \"kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2\") pod \"keystone-db-create-2bwq2\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.732695 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.738613 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wtzn8"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.744868 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.750177 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wtzn8"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.766451 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.766561 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzv4h\" (UniqueName: \"kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.767290 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.788039 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzv4h\" (UniqueName: \"kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h\") pod \"keystone-fbe8-account-create-update-2ngk5\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.844616 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.847002 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-ff26-account-create-update-xgdmp"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.848793 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.852924 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.862810 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ff26-account-create-update-xgdmp"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.869622 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.869826 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzhk9\" (UniqueName: \"kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.971543 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j47h\" (UniqueName: \"kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.971616 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.971754 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.971790 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzhk9\" (UniqueName: \"kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.972994 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.982533 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-gdkrq"] Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.984468 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.995065 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzhk9\" (UniqueName: \"kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9\") pod \"placement-db-create-wtzn8\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:37 crc kubenswrapper[4960]: I1201 20:00:37.995677 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gdkrq"] Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.071280 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.073961 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxftk\" (UniqueName: \"kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.074074 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j47h\" (UniqueName: \"kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.074098 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.074209 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.074988 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.081029 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5d22-account-create-update-mvkxq"] Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.085815 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.088765 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.093603 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j47h\" (UniqueName: \"kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h\") pod \"placement-ff26-account-create-update-xgdmp\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.094340 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5d22-account-create-update-mvkxq"] Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.169683 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.180459 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr77n\" (UniqueName: \"kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.180501 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.180553 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxftk\" (UniqueName: \"kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.180594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.181306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.196797 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxftk\" (UniqueName: \"kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk\") pod \"glance-db-create-gdkrq\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.282331 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr77n\" (UniqueName: \"kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.282384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.283163 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.305939 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr77n\" (UniqueName: \"kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n\") pod \"glance-5d22-account-create-update-mvkxq\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.341674 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.441412 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.445675 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"39fbd706-99ef-4b5d-9d7a-e84ee337c443","Type":"ContainerStarted","Data":"893215ed44b2da7e88f2e108f67834583c5adb28dc93d6ce452731d994683839"} Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.446535 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.450746 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.471805 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=33.184669081 podStartE2EDuration="59.471789774s" podCreationTimestamp="2025-12-01 19:59:39 +0000 UTC" firstStartedPulling="2025-12-01 20:00:07.753818377 +0000 UTC m=+1243.041310046" lastFinishedPulling="2025-12-01 20:00:34.04093907 +0000 UTC m=+1269.328430739" observedRunningTime="2025-12-01 20:00:38.466207459 +0000 UTC m=+1273.753699138" watchObservedRunningTime="2025-12-01 20:00:38.471789774 +0000 UTC m=+1273.759281443" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.568228 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 01 20:00:38 crc kubenswrapper[4960]: I1201 20:00:38.692821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:38 crc kubenswrapper[4960]: E1201 20:00:38.693020 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:38 crc kubenswrapper[4960]: E1201 20:00:38.693039 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:38 crc kubenswrapper[4960]: E1201 20:00:38.693083 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:00:46.693067159 +0000 UTC m=+1281.980558828 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:39 crc kubenswrapper[4960]: I1201 20:00:39.114771 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-distributor-56cd74f89f-jxqfj" Dec 01 20:00:39 crc kubenswrapper[4960]: I1201 20:00:39.405789 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-querier-548665d79b-gd6lz" Dec 01 20:00:39 crc kubenswrapper[4960]: I1201 20:00:39.459914 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-query-frontend-779849886d-hf7g4" Dec 01 20:00:39 crc kubenswrapper[4960]: I1201 20:00:39.515246 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.032506 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.273480 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="969241f0-8883-4570-94d4-fdb20ee094cd" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.340744 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-compactor-0" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.452988 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-index-gateway-0" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.471324 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.576849 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:40 crc kubenswrapper[4960]: I1201 20:00:40.587724 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="dnsmasq-dns" containerID="cri-o://6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61" gracePeriod=10 Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.316493 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.453262 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5d22-account-create-update-mvkxq"] Dec 01 20:00:41 crc kubenswrapper[4960]: W1201 20:00:41.456298 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dec74ac_5b50_42ed_9b0e_c58f881eef9c.slice/crio-bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad WatchSource:0}: Error finding container bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad: Status 404 returned error can't find the container with id bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.458794 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config\") pod \"001b0c13-5e8e-4458-839a-83059a9bd06f\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.458850 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb\") pod \"001b0c13-5e8e-4458-839a-83059a9bd06f\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.458920 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hnp4\" (UniqueName: \"kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4\") pod \"001b0c13-5e8e-4458-839a-83059a9bd06f\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.458980 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc\") pod \"001b0c13-5e8e-4458-839a-83059a9bd06f\" (UID: \"001b0c13-5e8e-4458-839a-83059a9bd06f\") " Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.467087 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4" (OuterVolumeSpecName: "kube-api-access-5hnp4") pod "001b0c13-5e8e-4458-839a-83059a9bd06f" (UID: "001b0c13-5e8e-4458-839a-83059a9bd06f"). InnerVolumeSpecName "kube-api-access-5hnp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.489052 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5d22-account-create-update-mvkxq" event={"ID":"6dec74ac-5b50-42ed-9b0e-c58f881eef9c","Type":"ContainerStarted","Data":"bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.493820 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerStarted","Data":"9ee4514d2a9ea5f7e45f991c76b1eaa965c4be86cd787f78ee2c9084242162fa"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.510458 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"07787692-57cc-473c-ab9b-276004fb9220","Type":"ContainerStarted","Data":"e4c0358b5849a5d6f29192c2385523b64e914ebf16633c1f07e97075079baf84"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.510515 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"07787692-57cc-473c-ab9b-276004fb9220","Type":"ContainerStarted","Data":"11ca0d7cda1fd4c14cd40767a60688265f16d3d895c7e1e8b4cc7c554d302f43"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.510567 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517173 4960 generic.go:334] "Generic (PLEG): container finished" podID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerID="6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61" exitCode=0 Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517221 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517263 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" event={"ID":"001b0c13-5e8e-4458-839a-83059a9bd06f","Type":"ContainerDied","Data":"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517295 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-t7hrw" event={"ID":"001b0c13-5e8e-4458-839a-83059a9bd06f","Type":"ContainerDied","Data":"2b8a05acb81b100790909277aaa3dd0dcfe368cdd3d1a9233d45e3e3c2d4bb7a"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517315 4960 scope.go:117] "RemoveContainer" containerID="6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.517723 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "001b0c13-5e8e-4458-839a-83059a9bd06f" (UID: "001b0c13-5e8e-4458-839a-83059a9bd06f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.519237 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmqjb" event={"ID":"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe","Type":"ContainerStarted","Data":"cff4b462cfe76812772e638f0b068e656b1ed6ba455a88be88593a59698be95a"} Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.530298 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "001b0c13-5e8e-4458-839a-83059a9bd06f" (UID: "001b0c13-5e8e-4458-839a-83059a9bd06f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.530470 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config" (OuterVolumeSpecName: "config") pod "001b0c13-5e8e-4458-839a-83059a9bd06f" (UID: "001b0c13-5e8e-4458-839a-83059a9bd06f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.548519 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.3303308449999998 podStartE2EDuration="10.548503897s" podCreationTimestamp="2025-12-01 20:00:31 +0000 UTC" firstStartedPulling="2025-12-01 20:00:32.555422455 +0000 UTC m=+1267.842914124" lastFinishedPulling="2025-12-01 20:00:40.773595507 +0000 UTC m=+1276.061087176" observedRunningTime="2025-12-01 20:00:41.536705479 +0000 UTC m=+1276.824197168" watchObservedRunningTime="2025-12-01 20:00:41.548503897 +0000 UTC m=+1276.835995566" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.555066 4960 scope.go:117] "RemoveContainer" containerID="fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.564005 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.564049 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.564065 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hnp4\" (UniqueName: \"kubernetes.io/projected/001b0c13-5e8e-4458-839a-83059a9bd06f-kube-api-access-5hnp4\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.564076 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/001b0c13-5e8e-4458-839a-83059a9bd06f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.565862 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-gmqjb" podStartSLOduration=2.126366621 podStartE2EDuration="10.565844149s" podCreationTimestamp="2025-12-01 20:00:31 +0000 UTC" firstStartedPulling="2025-12-01 20:00:32.446735233 +0000 UTC m=+1267.734226902" lastFinishedPulling="2025-12-01 20:00:40.886212761 +0000 UTC m=+1276.173704430" observedRunningTime="2025-12-01 20:00:41.562500132 +0000 UTC m=+1276.849991801" watchObservedRunningTime="2025-12-01 20:00:41.565844149 +0000 UTC m=+1276.853335818" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.599077 4960 scope.go:117] "RemoveContainer" containerID="6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61" Dec 01 20:00:41 crc kubenswrapper[4960]: E1201 20:00:41.599732 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61\": container with ID starting with 6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61 not found: ID does not exist" containerID="6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.599773 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61"} err="failed to get container status \"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61\": rpc error: code = NotFound desc = could not find container \"6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61\": container with ID starting with 6e34c029c86799333004d0cf763af84b58b4dc03afc23540c3d814cbe63ddd61 not found: ID does not exist" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.599804 4960 scope.go:117] "RemoveContainer" containerID="fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed" Dec 01 20:00:41 crc kubenswrapper[4960]: E1201 20:00:41.600210 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed\": container with ID starting with fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed not found: ID does not exist" containerID="fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.600236 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed"} err="failed to get container status \"fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed\": rpc error: code = NotFound desc = could not find container \"fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed\": container with ID starting with fe6dae7fb125200844fa8c6350f2bbf3f83b9b087438d93f30e4624eab13f3ed not found: ID does not exist" Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.819826 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-gdkrq"] Dec 01 20:00:41 crc kubenswrapper[4960]: W1201 20:00:41.820392 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce13c3c2_01f2_4d05_a8e8_3107617eaf29.slice/crio-83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530 WatchSource:0}: Error finding container 83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530: Status 404 returned error can't find the container with id 83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530 Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.843820 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fbe8-account-create-update-2ngk5"] Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.864829 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wtzn8"] Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.874169 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2bwq2"] Dec 01 20:00:41 crc kubenswrapper[4960]: I1201 20:00:41.881694 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-ff26-account-create-update-xgdmp"] Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.103242 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.113326 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-t7hrw"] Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.532010 4960 generic.go:334] "Generic (PLEG): container finished" podID="04fb1740-7860-4e08-9a62-13e55b59ed77" containerID="fff22dcc570d6e3492f343e709800b403f67337ea4d45d409d6c5bc1f472086e" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.532896 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wtzn8" event={"ID":"04fb1740-7860-4e08-9a62-13e55b59ed77","Type":"ContainerDied","Data":"fff22dcc570d6e3492f343e709800b403f67337ea4d45d409d6c5bc1f472086e"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.532930 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wtzn8" event={"ID":"04fb1740-7860-4e08-9a62-13e55b59ed77","Type":"ContainerStarted","Data":"708c930856c1a8c493673942493336c3425ac45550bb2b863448aeb3bfe5538b"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.538571 4960 generic.go:334] "Generic (PLEG): container finished" podID="13bd72fe-2556-4e33-bf2c-f95eba6d2310" containerID="46ccfc853f030042a7463e7ac6d328af94ad03c4a4a753919a7ed53a294dbe62" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.538657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ff26-account-create-update-xgdmp" event={"ID":"13bd72fe-2556-4e33-bf2c-f95eba6d2310","Type":"ContainerDied","Data":"46ccfc853f030042a7463e7ac6d328af94ad03c4a4a753919a7ed53a294dbe62"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.538698 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ff26-account-create-update-xgdmp" event={"ID":"13bd72fe-2556-4e33-bf2c-f95eba6d2310","Type":"ContainerStarted","Data":"1a66491925746059bfb9b67f1ca3b97652111654d7ac87bb477aca0aefa5d10b"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.540333 4960 generic.go:334] "Generic (PLEG): container finished" podID="ec2068f8-377d-454f-80f3-429d30d4ecaa" containerID="7a9c7dd926d087fc72804d3ce9caf7d55a31fbeb6e47935be41f7689ff48ab6f" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.540404 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2bwq2" event={"ID":"ec2068f8-377d-454f-80f3-429d30d4ecaa","Type":"ContainerDied","Data":"7a9c7dd926d087fc72804d3ce9caf7d55a31fbeb6e47935be41f7689ff48ab6f"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.540421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2bwq2" event={"ID":"ec2068f8-377d-454f-80f3-429d30d4ecaa","Type":"ContainerStarted","Data":"dc498e605a0ed769ea45538121dc3759c6f808ce9ce5eac3440c876c12b9aeac"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.542415 4960 generic.go:334] "Generic (PLEG): container finished" podID="51c22ac6-d753-4723-8889-c839eb542bf1" containerID="dc30df6ecc7121cad23b772db529e2a007209ede885b3c396bdf0c333936f0b6" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.542471 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fbe8-account-create-update-2ngk5" event={"ID":"51c22ac6-d753-4723-8889-c839eb542bf1","Type":"ContainerDied","Data":"dc30df6ecc7121cad23b772db529e2a007209ede885b3c396bdf0c333936f0b6"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.542486 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fbe8-account-create-update-2ngk5" event={"ID":"51c22ac6-d753-4723-8889-c839eb542bf1","Type":"ContainerStarted","Data":"06a6bbebca1649eef36bf99978e39dc02f50896dd6c6a842afaadbca0e8880e6"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.545410 4960 generic.go:334] "Generic (PLEG): container finished" podID="6dec74ac-5b50-42ed-9b0e-c58f881eef9c" containerID="2c3189f656ce75428367a615a03c7efc9c7e8ae0409bb7d370ac2a80de90ca2a" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.545471 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5d22-account-create-update-mvkxq" event={"ID":"6dec74ac-5b50-42ed-9b0e-c58f881eef9c","Type":"ContainerDied","Data":"2c3189f656ce75428367a615a03c7efc9c7e8ae0409bb7d370ac2a80de90ca2a"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.549727 4960 generic.go:334] "Generic (PLEG): container finished" podID="ce13c3c2-01f2-4d05-a8e8-3107617eaf29" containerID="f55900b7072a38b367473834cf0325067c71eadd1305de7e1f6b8fdd4834a91b" exitCode=0 Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.549790 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gdkrq" event={"ID":"ce13c3c2-01f2-4d05-a8e8-3107617eaf29","Type":"ContainerDied","Data":"f55900b7072a38b367473834cf0325067c71eadd1305de7e1f6b8fdd4834a91b"} Dec 01 20:00:42 crc kubenswrapper[4960]: I1201 20:00:42.549840 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gdkrq" event={"ID":"ce13c3c2-01f2-4d05-a8e8-3107617eaf29","Type":"ContainerStarted","Data":"83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530"} Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.337173 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" path="/var/lib/kubelet/pods/001b0c13-5e8e-4458-839a-83059a9bd06f/volumes" Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.871156 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.871212 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.871263 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.872016 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:00:43 crc kubenswrapper[4960]: I1201 20:00:43.872061 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07" gracePeriod=600 Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.178404 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.329127 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts\") pod \"51c22ac6-d753-4723-8889-c839eb542bf1\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.329176 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzv4h\" (UniqueName: \"kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h\") pod \"51c22ac6-d753-4723-8889-c839eb542bf1\" (UID: \"51c22ac6-d753-4723-8889-c839eb542bf1\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.330244 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51c22ac6-d753-4723-8889-c839eb542bf1" (UID: "51c22ac6-d753-4723-8889-c839eb542bf1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.349030 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h" (OuterVolumeSpecName: "kube-api-access-pzv4h") pod "51c22ac6-d753-4723-8889-c839eb542bf1" (UID: "51c22ac6-d753-4723-8889-c839eb542bf1"). InnerVolumeSpecName "kube-api-access-pzv4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.431674 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51c22ac6-d753-4723-8889-c839eb542bf1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.431724 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzv4h\" (UniqueName: \"kubernetes.io/projected/51c22ac6-d753-4723-8889-c839eb542bf1-kube-api-access-pzv4h\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.468043 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.480410 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.497835 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.511483 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.528889 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.532619 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts\") pod \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.532823 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr77n\" (UniqueName: \"kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n\") pod \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\" (UID: \"6dec74ac-5b50-42ed-9b0e-c58f881eef9c\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.534198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dec74ac-5b50-42ed-9b0e-c58f881eef9c" (UID: "6dec74ac-5b50-42ed-9b0e-c58f881eef9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.543972 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n" (OuterVolumeSpecName: "kube-api-access-qr77n") pod "6dec74ac-5b50-42ed-9b0e-c58f881eef9c" (UID: "6dec74ac-5b50-42ed-9b0e-c58f881eef9c"). InnerVolumeSpecName "kube-api-access-qr77n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.580854 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-gdkrq" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.580889 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-gdkrq" event={"ID":"ce13c3c2-01f2-4d05-a8e8-3107617eaf29","Type":"ContainerDied","Data":"83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.580933 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83c5435e0dda48fde71ba72688958a10fdea1bff107db39b9823b7cc6e087530" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.587200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerStarted","Data":"d004289dbdeb01bc50e9cbc5bd03e9e15b481c8815376315d28c611c6671b2fb"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.596543 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wtzn8" event={"ID":"04fb1740-7860-4e08-9a62-13e55b59ed77","Type":"ContainerDied","Data":"708c930856c1a8c493673942493336c3425ac45550bb2b863448aeb3bfe5538b"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.596590 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="708c930856c1a8c493673942493336c3425ac45550bb2b863448aeb3bfe5538b" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.596659 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wtzn8" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.599155 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-ff26-account-create-update-xgdmp" event={"ID":"13bd72fe-2556-4e33-bf2c-f95eba6d2310","Type":"ContainerDied","Data":"1a66491925746059bfb9b67f1ca3b97652111654d7ac87bb477aca0aefa5d10b"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.599204 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a66491925746059bfb9b67f1ca3b97652111654d7ac87bb477aca0aefa5d10b" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.599255 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-ff26-account-create-update-xgdmp" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.606937 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07" exitCode=0 Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.607034 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.607088 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.607105 4960 scope.go:117] "RemoveContainer" containerID="e42d643895e7adc8b1aa5e6747fc8aadf6c95bb4f21a6cfbf8c41a221813ba09" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.609535 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2bwq2" event={"ID":"ec2068f8-377d-454f-80f3-429d30d4ecaa","Type":"ContainerDied","Data":"dc498e605a0ed769ea45538121dc3759c6f808ce9ce5eac3440c876c12b9aeac"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.609556 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc498e605a0ed769ea45538121dc3759c6f808ce9ce5eac3440c876c12b9aeac" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.609608 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2bwq2" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.634934 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j47h\" (UniqueName: \"kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h\") pod \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.635065 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts\") pod \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.635143 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26dt2\" (UniqueName: \"kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2\") pod \"ec2068f8-377d-454f-80f3-429d30d4ecaa\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.635191 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzhk9\" (UniqueName: \"kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9\") pod \"04fb1740-7860-4e08-9a62-13e55b59ed77\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.639591 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fbe8-account-create-update-2ngk5" event={"ID":"51c22ac6-d753-4723-8889-c839eb542bf1","Type":"ContainerDied","Data":"06a6bbebca1649eef36bf99978e39dc02f50896dd6c6a842afaadbca0e8880e6"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.639665 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06a6bbebca1649eef36bf99978e39dc02f50896dd6c6a842afaadbca0e8880e6" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.639753 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fbe8-account-create-update-2ngk5" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.640469 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce13c3c2-01f2-4d05-a8e8-3107617eaf29" (UID: "ce13c3c2-01f2-4d05-a8e8-3107617eaf29"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.644566 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts\") pod \"ec2068f8-377d-454f-80f3-429d30d4ecaa\" (UID: \"ec2068f8-377d-454f-80f3-429d30d4ecaa\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.644644 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxftk\" (UniqueName: \"kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk\") pod \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\" (UID: \"ce13c3c2-01f2-4d05-a8e8-3107617eaf29\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.644739 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts\") pod \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\" (UID: \"13bd72fe-2556-4e33-bf2c-f95eba6d2310\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.644873 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts\") pod \"04fb1740-7860-4e08-9a62-13e55b59ed77\" (UID: \"04fb1740-7860-4e08-9a62-13e55b59ed77\") " Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.645995 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9" (OuterVolumeSpecName: "kube-api-access-xzhk9") pod "04fb1740-7860-4e08-9a62-13e55b59ed77" (UID: "04fb1740-7860-4e08-9a62-13e55b59ed77"). InnerVolumeSpecName "kube-api-access-xzhk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.646602 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "13bd72fe-2556-4e33-bf2c-f95eba6d2310" (UID: "13bd72fe-2556-4e33-bf2c-f95eba6d2310"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.647359 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04fb1740-7860-4e08-9a62-13e55b59ed77" (UID: "04fb1740-7860-4e08-9a62-13e55b59ed77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.648155 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5d22-account-create-update-mvkxq" event={"ID":"6dec74ac-5b50-42ed-9b0e-c58f881eef9c","Type":"ContainerDied","Data":"bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad"} Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.648202 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc4444538bc62e55fa187c2734c5c42030a1a0698a6b7a9b43732f48463b7bad" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.648304 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5d22-account-create-update-mvkxq" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.649904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2" (OuterVolumeSpecName: "kube-api-access-26dt2") pod "ec2068f8-377d-454f-80f3-429d30d4ecaa" (UID: "ec2068f8-377d-454f-80f3-429d30d4ecaa"). InnerVolumeSpecName "kube-api-access-26dt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.651600 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec2068f8-377d-454f-80f3-429d30d4ecaa" (UID: "ec2068f8-377d-454f-80f3-429d30d4ecaa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653483 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13bd72fe-2556-4e33-bf2c-f95eba6d2310-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653516 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04fb1740-7860-4e08-9a62-13e55b59ed77-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653532 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653558 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653576 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26dt2\" (UniqueName: \"kubernetes.io/projected/ec2068f8-377d-454f-80f3-429d30d4ecaa-kube-api-access-26dt2\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653591 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzhk9\" (UniqueName: \"kubernetes.io/projected/04fb1740-7860-4e08-9a62-13e55b59ed77-kube-api-access-xzhk9\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.653609 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr77n\" (UniqueName: \"kubernetes.io/projected/6dec74ac-5b50-42ed-9b0e-c58f881eef9c-kube-api-access-qr77n\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.654265 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk" (OuterVolumeSpecName: "kube-api-access-dxftk") pod "ce13c3c2-01f2-4d05-a8e8-3107617eaf29" (UID: "ce13c3c2-01f2-4d05-a8e8-3107617eaf29"). InnerVolumeSpecName "kube-api-access-dxftk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.654684 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h" (OuterVolumeSpecName: "kube-api-access-2j47h") pod "13bd72fe-2556-4e33-bf2c-f95eba6d2310" (UID: "13bd72fe-2556-4e33-bf2c-f95eba6d2310"). InnerVolumeSpecName "kube-api-access-2j47h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.756131 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j47h\" (UniqueName: \"kubernetes.io/projected/13bd72fe-2556-4e33-bf2c-f95eba6d2310-kube-api-access-2j47h\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.756182 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec2068f8-377d-454f-80f3-429d30d4ecaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:44 crc kubenswrapper[4960]: I1201 20:00:44.756195 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxftk\" (UniqueName: \"kubernetes.io/projected/ce13c3c2-01f2-4d05-a8e8-3107617eaf29-kube-api-access-dxftk\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:46 crc kubenswrapper[4960]: I1201 20:00:46.794866 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:00:46 crc kubenswrapper[4960]: E1201 20:00:46.795102 4960 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 01 20:00:46 crc kubenswrapper[4960]: E1201 20:00:46.795510 4960 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 01 20:00:46 crc kubenswrapper[4960]: E1201 20:00:46.795562 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift podName:725698af-f23c-4262-96a4-2bfff12018fa nodeName:}" failed. No retries permitted until 2025-12-01 20:01:02.795543747 +0000 UTC m=+1298.083035416 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift") pod "swift-storage-0" (UID: "725698af-f23c-4262-96a4-2bfff12018fa") : configmap "swift-ring-files" not found Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.341614 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-r6j4l"] Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342454 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c22ac6-d753-4723-8889-c839eb542bf1" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342471 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c22ac6-d753-4723-8889-c839eb542bf1" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342484 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13bd72fe-2556-4e33-bf2c-f95eba6d2310" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342490 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="13bd72fe-2556-4e33-bf2c-f95eba6d2310" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342507 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fb1740-7860-4e08-9a62-13e55b59ed77" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342513 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fb1740-7860-4e08-9a62-13e55b59ed77" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342532 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec2068f8-377d-454f-80f3-429d30d4ecaa" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342538 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec2068f8-377d-454f-80f3-429d30d4ecaa" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342547 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dec74ac-5b50-42ed-9b0e-c58f881eef9c" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342553 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dec74ac-5b50-42ed-9b0e-c58f881eef9c" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342560 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="init" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342565 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="init" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342579 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce13c3c2-01f2-4d05-a8e8-3107617eaf29" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342584 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce13c3c2-01f2-4d05-a8e8-3107617eaf29" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: E1201 20:00:48.342597 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="dnsmasq-dns" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342603 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="dnsmasq-dns" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342802 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="001b0c13-5e8e-4458-839a-83059a9bd06f" containerName="dnsmasq-dns" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342817 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="04fb1740-7860-4e08-9a62-13e55b59ed77" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342833 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dec74ac-5b50-42ed-9b0e-c58f881eef9c" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342841 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec2068f8-377d-454f-80f3-429d30d4ecaa" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342850 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce13c3c2-01f2-4d05-a8e8-3107617eaf29" containerName="mariadb-database-create" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342858 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c22ac6-d753-4723-8889-c839eb542bf1" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.342870 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="13bd72fe-2556-4e33-bf2c-f95eba6d2310" containerName="mariadb-account-create-update" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.343597 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.346229 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mb8dw" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.346267 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.350016 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r6j4l"] Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.432384 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.432462 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.432644 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlk9g\" (UniqueName: \"kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.433178 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.534562 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlk9g\" (UniqueName: \"kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.534689 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.534752 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.534785 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.549393 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.549680 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.551005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.554326 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlk9g\" (UniqueName: \"kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g\") pod \"glance-db-sync-r6j4l\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:48 crc kubenswrapper[4960]: I1201 20:00:48.669237 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r6j4l" Dec 01 20:00:49 crc kubenswrapper[4960]: I1201 20:00:49.283940 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrf7z" podUID="87e16436-45f5-4392-b524-b86f1cc5c0b2" containerName="ovn-controller" probeResult="failure" output=< Dec 01 20:00:49 crc kubenswrapper[4960]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 20:00:49 crc kubenswrapper[4960]: > Dec 01 20:00:49 crc kubenswrapper[4960]: I1201 20:00:49.699783 4960 generic.go:334] "Generic (PLEG): container finished" podID="aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" containerID="cff4b462cfe76812772e638f0b068e656b1ed6ba455a88be88593a59698be95a" exitCode=0 Dec 01 20:00:49 crc kubenswrapper[4960]: I1201 20:00:49.699843 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmqjb" event={"ID":"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe","Type":"ContainerDied","Data":"cff4b462cfe76812772e638f0b068e656b1ed6ba455a88be88593a59698be95a"} Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.235976 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r6j4l"] Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.273637 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="969241f0-8883-4570-94d4-fdb20ee094cd" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.712375 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r6j4l" event={"ID":"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec","Type":"ContainerStarted","Data":"32503d58ab175a217b9cb1e5ad97e07cdeb062eaa14a56b4d4663e9309ffa78b"} Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.722887 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerStarted","Data":"b859c505d25ebfa473024a4846702eaaae3b5604281547b039c6ce6d256ccfee"} Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.766033 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=28.832933698 podStartE2EDuration="1m11.766011176s" podCreationTimestamp="2025-12-01 19:59:39 +0000 UTC" firstStartedPulling="2025-12-01 20:00:06.802331196 +0000 UTC m=+1242.089822865" lastFinishedPulling="2025-12-01 20:00:49.735408664 +0000 UTC m=+1285.022900343" observedRunningTime="2025-12-01 20:00:50.755077191 +0000 UTC m=+1286.042568920" watchObservedRunningTime="2025-12-01 20:00:50.766011176 +0000 UTC m=+1286.053502855" Dec 01 20:00:50 crc kubenswrapper[4960]: I1201 20:00:50.835825 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.214414 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.391708 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.391764 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.391806 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.391872 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.391944 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.392066 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.392588 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.393157 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.393593 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rtvx\" (UniqueName: \"kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx\") pod \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\" (UID: \"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe\") " Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.394305 4960 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.394331 4960 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.406005 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx" (OuterVolumeSpecName: "kube-api-access-9rtvx") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "kube-api-access-9rtvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.410130 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.423154 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.425312 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.442821 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts" (OuterVolumeSpecName: "scripts") pod "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" (UID: "aa3614b7-1a8b-4007-9e01-f6340ff9b0fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.496188 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.496281 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rtvx\" (UniqueName: \"kubernetes.io/projected/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-kube-api-access-9rtvx\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.496299 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.496340 4960 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.496355 4960 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aa3614b7-1a8b-4007-9e01-f6340ff9b0fe-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.733196 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gmqjb" event={"ID":"aa3614b7-1a8b-4007-9e01-f6340ff9b0fe","Type":"ContainerDied","Data":"98dfe4797cf0534f9d66b999e38dbd5c8e851436226a4ce01eb4b156fe6ee504"} Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.733238 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98dfe4797cf0534f9d66b999e38dbd5c8e851436226a4ce01eb4b156fe6ee504" Dec 01 20:00:51 crc kubenswrapper[4960]: I1201 20:00:51.733274 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gmqjb" Dec 01 20:00:52 crc kubenswrapper[4960]: I1201 20:00:52.019673 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 01 20:00:54 crc kubenswrapper[4960]: I1201 20:00:54.293532 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrf7z" podUID="87e16436-45f5-4392-b524-b86f1cc5c0b2" containerName="ovn-controller" probeResult="failure" output=< Dec 01 20:00:54 crc kubenswrapper[4960]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 20:00:54 crc kubenswrapper[4960]: > Dec 01 20:00:54 crc kubenswrapper[4960]: I1201 20:00:54.313890 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.774720 4960 generic.go:334] "Generic (PLEG): container finished" podID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerID="091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577" exitCode=0 Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.774834 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerDied","Data":"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577"} Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.777697 4960 generic.go:334] "Generic (PLEG): container finished" podID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerID="d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5" exitCode=0 Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.777741 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerDied","Data":"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5"} Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.835527 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 20:00:55 crc kubenswrapper[4960]: I1201 20:00:55.840507 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 20:00:56 crc kubenswrapper[4960]: I1201 20:00:56.789409 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.282123 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-nrf7z" podUID="87e16436-45f5-4392-b524-b86f1cc5c0b2" containerName="ovn-controller" probeResult="failure" output=< Dec 01 20:00:59 crc kubenswrapper[4960]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 01 20:00:59 crc kubenswrapper[4960]: > Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.312681 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-9n6l4" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.430506 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.430849 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="prometheus" containerID="cri-o://9ee4514d2a9ea5f7e45f991c76b1eaa965c4be86cd787f78ee2c9084242162fa" gracePeriod=600 Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.431016 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="thanos-sidecar" containerID="cri-o://b859c505d25ebfa473024a4846702eaaae3b5604281547b039c6ce6d256ccfee" gracePeriod=600 Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.431082 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="config-reloader" containerID="cri-o://d004289dbdeb01bc50e9cbc5bd03e9e15b481c8815376315d28c611c6671b2fb" gracePeriod=600 Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.538512 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-nrf7z-config-bg7wm"] Dec 01 20:00:59 crc kubenswrapper[4960]: E1201 20:00:59.539028 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" containerName="swift-ring-rebalance" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.539055 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" containerName="swift-ring-rebalance" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.540512 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa3614b7-1a8b-4007-9e01-f6340ff9b0fe" containerName="swift-ring-rebalance" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.541495 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.554366 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.558021 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrf7z-config-bg7wm"] Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.679797 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.679865 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.679915 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.679998 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx9ml\" (UniqueName: \"kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.680019 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.680034 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781164 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx9ml\" (UniqueName: \"kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781222 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781241 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781288 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781323 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781689 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781752 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.781785 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.782306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.785695 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.812135 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx9ml\" (UniqueName: \"kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml\") pod \"ovn-controller-nrf7z-config-bg7wm\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:00:59 crc kubenswrapper[4960]: I1201 20:00:59.888834 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:01:00 crc kubenswrapper[4960]: I1201 20:01:00.271245 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="969241f0-8883-4570-94d4-fdb20ee094cd" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 20:01:00 crc kubenswrapper[4960]: I1201 20:01:00.837384 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="prometheus" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831033 4960 generic.go:334] "Generic (PLEG): container finished" podID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerID="b859c505d25ebfa473024a4846702eaaae3b5604281547b039c6ce6d256ccfee" exitCode=0 Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831073 4960 generic.go:334] "Generic (PLEG): container finished" podID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerID="d004289dbdeb01bc50e9cbc5bd03e9e15b481c8815376315d28c611c6671b2fb" exitCode=0 Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831083 4960 generic.go:334] "Generic (PLEG): container finished" podID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerID="9ee4514d2a9ea5f7e45f991c76b1eaa965c4be86cd787f78ee2c9084242162fa" exitCode=0 Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831159 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerDied","Data":"b859c505d25ebfa473024a4846702eaaae3b5604281547b039c6ce6d256ccfee"} Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831204 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerDied","Data":"d004289dbdeb01bc50e9cbc5bd03e9e15b481c8815376315d28c611c6671b2fb"} Dec 01 20:01:01 crc kubenswrapper[4960]: I1201 20:01:01.831217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerDied","Data":"9ee4514d2a9ea5f7e45f991c76b1eaa965c4be86cd787f78ee2c9084242162fa"} Dec 01 20:01:02 crc kubenswrapper[4960]: I1201 20:01:02.843061 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:01:02 crc kubenswrapper[4960]: I1201 20:01:02.851402 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/725698af-f23c-4262-96a4-2bfff12018fa-etc-swift\") pod \"swift-storage-0\" (UID: \"725698af-f23c-4262-96a4-2bfff12018fa\") " pod="openstack/swift-storage-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.092843 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.234603 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.356951 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357037 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357081 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357308 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357353 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357404 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357493 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmtcn\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.357521 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config\") pod \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\" (UID: \"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7\") " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.358434 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.362747 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.364393 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out" (OuterVolumeSpecName: "config-out") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.365371 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.365821 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn" (OuterVolumeSpecName: "kube-api-access-rmtcn") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "kube-api-access-rmtcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.368349 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config" (OuterVolumeSpecName: "config") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.395444 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config" (OuterVolumeSpecName: "web-config") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.403135 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" (UID: "92edfae7-7d6c-4f89-a54d-c9f2c2d422f7"). InnerVolumeSpecName "pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.441317 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-nrf7z-config-bg7wm"] Dec 01 20:01:03 crc kubenswrapper[4960]: W1201 20:01:03.441737 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eea7d45_e33e_4bec_9810_bee3cff9c612.slice/crio-8c75feae0b76285ad085f1e0a63d6e97e914f7f5f96cb4d0eeaae8fd8096562b WatchSource:0}: Error finding container 8c75feae0b76285ad085f1e0a63d6e97e914f7f5f96cb4d0eeaae8fd8096562b: Status 404 returned error can't find the container with id 8c75feae0b76285ad085f1e0a63d6e97e914f7f5f96cb4d0eeaae8fd8096562b Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464647 4960 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config-out\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464705 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") on node \"crc\" " Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464719 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464729 4960 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464739 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmtcn\" (UniqueName: \"kubernetes.io/projected/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-kube-api-access-rmtcn\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464748 4960 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-web-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464757 4960 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.464765 4960 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.487215 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.487363 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c") on node "crc" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.565951 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.702919 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 01 20:01:03 crc kubenswrapper[4960]: W1201 20:01:03.711833 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod725698af_f23c_4262_96a4_2bfff12018fa.slice/crio-e87ad35d72a7132624242a2709a16e200d194b6416ab9573d08b72bb3ad1da34 WatchSource:0}: Error finding container e87ad35d72a7132624242a2709a16e200d194b6416ab9573d08b72bb3ad1da34: Status 404 returned error can't find the container with id e87ad35d72a7132624242a2709a16e200d194b6416ab9573d08b72bb3ad1da34 Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.848944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r6j4l" event={"ID":"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec","Type":"ContainerStarted","Data":"7372ee4e45a5f3cce7e2ff786d7314531acc1f0d8351fb1e1582def7a5cbc28a"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.851673 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.851667 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92edfae7-7d6c-4f89-a54d-c9f2c2d422f7","Type":"ContainerDied","Data":"02b7df7b8aaa82675031f62eb345d619b02e6ad789da191811b50770022c806c"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.851794 4960 scope.go:117] "RemoveContainer" containerID="b859c505d25ebfa473024a4846702eaaae3b5604281547b039c6ce6d256ccfee" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.853518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrf7z-config-bg7wm" event={"ID":"6eea7d45-e33e-4bec-9810-bee3cff9c612","Type":"ContainerStarted","Data":"8a228de05fbd4b29f1cbedc4d66e6a2ca34dccefd0c3c14d02cacaaa55c5270b"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.853569 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrf7z-config-bg7wm" event={"ID":"6eea7d45-e33e-4bec-9810-bee3cff9c612","Type":"ContainerStarted","Data":"8c75feae0b76285ad085f1e0a63d6e97e914f7f5f96cb4d0eeaae8fd8096562b"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.856737 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerStarted","Data":"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.857021 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.859823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerStarted","Data":"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.860785 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.863828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"e87ad35d72a7132624242a2709a16e200d194b6416ab9573d08b72bb3ad1da34"} Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.874694 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-r6j4l" podStartSLOduration=3.138051993 podStartE2EDuration="15.874647313s" podCreationTimestamp="2025-12-01 20:00:48 +0000 UTC" firstStartedPulling="2025-12-01 20:00:50.245458773 +0000 UTC m=+1285.532950442" lastFinishedPulling="2025-12-01 20:01:02.982054083 +0000 UTC m=+1298.269545762" observedRunningTime="2025-12-01 20:01:03.869220975 +0000 UTC m=+1299.156712644" watchObservedRunningTime="2025-12-01 20:01:03.874647313 +0000 UTC m=+1299.162138982" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.891008 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-nrf7z-config-bg7wm" podStartSLOduration=4.890990608 podStartE2EDuration="4.890990608s" podCreationTimestamp="2025-12-01 20:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:03.88876765 +0000 UTC m=+1299.176259309" watchObservedRunningTime="2025-12-01 20:01:03.890990608 +0000 UTC m=+1299.178482267" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.919382 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.79169912 podStartE2EDuration="1m31.919359137s" podCreationTimestamp="2025-12-01 19:59:32 +0000 UTC" firstStartedPulling="2025-12-01 19:59:35.020531981 +0000 UTC m=+1210.308023650" lastFinishedPulling="2025-12-01 20:00:18.148191988 +0000 UTC m=+1253.435683667" observedRunningTime="2025-12-01 20:01:03.916459867 +0000 UTC m=+1299.203951556" watchObservedRunningTime="2025-12-01 20:01:03.919359137 +0000 UTC m=+1299.206850806" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.929070 4960 scope.go:117] "RemoveContainer" containerID="d004289dbdeb01bc50e9cbc5bd03e9e15b481c8815376315d28c611c6671b2fb" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.949401 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=48.764425686 podStartE2EDuration="1m31.949384166s" podCreationTimestamp="2025-12-01 19:59:32 +0000 UTC" firstStartedPulling="2025-12-01 19:59:34.853342106 +0000 UTC m=+1210.140833775" lastFinishedPulling="2025-12-01 20:00:18.038300586 +0000 UTC m=+1253.325792255" observedRunningTime="2025-12-01 20:01:03.946036012 +0000 UTC m=+1299.233527691" watchObservedRunningTime="2025-12-01 20:01:03.949384166 +0000 UTC m=+1299.236875835" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.950057 4960 scope.go:117] "RemoveContainer" containerID="9ee4514d2a9ea5f7e45f991c76b1eaa965c4be86cd787f78ee2c9084242162fa" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.968635 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.974782 4960 scope.go:117] "RemoveContainer" containerID="277493da9f058effbde9f62ad5cd02fcd55a6dbc036097577f98bd4b1b9a3ed9" Dec 01 20:01:03 crc kubenswrapper[4960]: I1201 20:01:03.976906 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.016461 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:01:04 crc kubenswrapper[4960]: E1201 20:01:04.016915 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="config-reloader" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.016930 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="config-reloader" Dec 01 20:01:04 crc kubenswrapper[4960]: E1201 20:01:04.016961 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="thanos-sidecar" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.016969 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="thanos-sidecar" Dec 01 20:01:04 crc kubenswrapper[4960]: E1201 20:01:04.016987 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="init-config-reloader" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.016996 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="init-config-reloader" Dec 01 20:01:04 crc kubenswrapper[4960]: E1201 20:01:04.017011 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="prometheus" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.017018 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="prometheus" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.017276 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="thanos-sidecar" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.017292 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="prometheus" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.017309 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" containerName="config-reloader" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.020683 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.023609 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.024275 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.024293 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-lkllm" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.024416 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.024464 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.029690 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.040928 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.047982 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074442 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074508 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw2sm\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-kube-api-access-rw2sm\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074567 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074727 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074817 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074854 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9a4f20da-f7ef-4710-9222-77baad783676-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.074887 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.075055 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.075090 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.075177 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9a4f20da-f7ef-4710-9222-77baad783676-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176684 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9a4f20da-f7ef-4710-9222-77baad783676-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176736 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176786 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw2sm\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-kube-api-access-rw2sm\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176844 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176877 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176905 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176925 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9a4f20da-f7ef-4710-9222-77baad783676-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176941 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.176999 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.177017 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.177490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9a4f20da-f7ef-4710-9222-77baad783676-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.183509 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.184796 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.184937 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.184959 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/aa508db444fabf42cff0f8d4f7fdb19d89c1265d06a9375978c8e8544e26865b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.185938 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9a4f20da-f7ef-4710-9222-77baad783676-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.192327 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.192813 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.193342 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.210511 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.210731 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/9a4f20da-f7ef-4710-9222-77baad783676-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.227864 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw2sm\" (UniqueName: \"kubernetes.io/projected/9a4f20da-f7ef-4710-9222-77baad783676-kube-api-access-rw2sm\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.288533 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-nrf7z" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.296422 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c3d40c94-3e8d-476f-9a26-57fb3c83034c\") pod \"prometheus-metric-storage-0\" (UID: \"9a4f20da-f7ef-4710-9222-77baad783676\") " pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.341473 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.884760 4960 generic.go:334] "Generic (PLEG): container finished" podID="6eea7d45-e33e-4bec-9810-bee3cff9c612" containerID="8a228de05fbd4b29f1cbedc4d66e6a2ca34dccefd0c3c14d02cacaaa55c5270b" exitCode=0 Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.885155 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-nrf7z-config-bg7wm" event={"ID":"6eea7d45-e33e-4bec-9810-bee3cff9c612","Type":"ContainerDied","Data":"8a228de05fbd4b29f1cbedc4d66e6a2ca34dccefd0c3c14d02cacaaa55c5270b"} Dec 01 20:01:04 crc kubenswrapper[4960]: I1201 20:01:04.917598 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 01 20:01:05 crc kubenswrapper[4960]: I1201 20:01:05.339181 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92edfae7-7d6c-4f89-a54d-c9f2c2d422f7" path="/var/lib/kubelet/pods/92edfae7-7d6c-4f89-a54d-c9f2c2d422f7/volumes" Dec 01 20:01:05 crc kubenswrapper[4960]: I1201 20:01:05.900623 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerStarted","Data":"beff73bc30057603db225a59fa16aad76dd52155190c8d51af9881ec91a48a71"} Dec 01 20:01:05 crc kubenswrapper[4960]: I1201 20:01:05.901576 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"e0e82e3d8161ed298e5e2a7cca66931d4cdc40ca7d1fc8b9312660a0a579d0e8"} Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.275558 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320052 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320092 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320142 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320228 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320233 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320446 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320476 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx9ml\" (UniqueName: \"kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml\") pod \"6eea7d45-e33e-4bec-9810-bee3cff9c612\" (UID: \"6eea7d45-e33e-4bec-9810-bee3cff9c612\") " Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320595 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.320670 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run" (OuterVolumeSpecName: "var-run") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.321213 4960 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.321234 4960 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.321245 4960 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6eea7d45-e33e-4bec-9810-bee3cff9c612-var-run\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.321399 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.321597 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts" (OuterVolumeSpecName: "scripts") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.326617 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml" (OuterVolumeSpecName: "kube-api-access-lx9ml") pod "6eea7d45-e33e-4bec-9810-bee3cff9c612" (UID: "6eea7d45-e33e-4bec-9810-bee3cff9c612"). InnerVolumeSpecName "kube-api-access-lx9ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.422738 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx9ml\" (UniqueName: \"kubernetes.io/projected/6eea7d45-e33e-4bec-9810-bee3cff9c612-kube-api-access-lx9ml\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.422776 4960 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.422786 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6eea7d45-e33e-4bec-9810-bee3cff9c612-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.530451 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-nrf7z-config-bg7wm"] Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.536167 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-nrf7z-config-bg7wm"] Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.914529 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-nrf7z-config-bg7wm" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.914698 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c75feae0b76285ad085f1e0a63d6e97e914f7f5f96cb4d0eeaae8fd8096562b" Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.920340 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"e9717f9da0c395077713af7e9c285620168aa50bf7a5be9f5e9f5328b009eb18"} Dec 01 20:01:06 crc kubenswrapper[4960]: I1201 20:01:06.920468 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"19d463809502ef42c081644b8d114b6c3781f49db49a10176d47a0154d8a1a9a"} Dec 01 20:01:07 crc kubenswrapper[4960]: I1201 20:01:07.343423 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eea7d45-e33e-4bec-9810-bee3cff9c612" path="/var/lib/kubelet/pods/6eea7d45-e33e-4bec-9810-bee3cff9c612/volumes" Dec 01 20:01:07 crc kubenswrapper[4960]: I1201 20:01:07.932997 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"31ff5bb27fff0749d27a3822f070050efe3a1b4f7427b6f0e5a99a3ff0fefe6d"} Dec 01 20:01:08 crc kubenswrapper[4960]: I1201 20:01:08.945715 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerStarted","Data":"3bb62657767591097ba8f70da2c3f0e435069c20d5903734eaf578dbca202f99"} Dec 01 20:01:09 crc kubenswrapper[4960]: I1201 20:01:09.956806 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"ffa9f389c9bd38d3265632466cb9aa4573e03243bc27a0f7226fd50272eea6c0"} Dec 01 20:01:09 crc kubenswrapper[4960]: I1201 20:01:09.957404 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"0d57047f7d1a4e2be90cb90627367d8e9be9b4d7d2a0d72baa040ec9977ddaa6"} Dec 01 20:01:09 crc kubenswrapper[4960]: I1201 20:01:09.957421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"7b870a9f4608001873302a38ed01a99c1a467e3c447f7c2a22b022de51bb43a1"} Dec 01 20:01:09 crc kubenswrapper[4960]: I1201 20:01:09.957432 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"b8b031fc4aedf1f0b336af4e1dabec6e1627cf7503dafd853b470909f8e90814"} Dec 01 20:01:10 crc kubenswrapper[4960]: I1201 20:01:10.269747 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cloudkitty-lokistack-ingester-0" podUID="969241f0-8883-4570-94d4-fdb20ee094cd" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 01 20:01:10 crc kubenswrapper[4960]: I1201 20:01:10.967375 4960 generic.go:334] "Generic (PLEG): container finished" podID="ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" containerID="7372ee4e45a5f3cce7e2ff786d7314531acc1f0d8351fb1e1582def7a5cbc28a" exitCode=0 Dec 01 20:01:10 crc kubenswrapper[4960]: I1201 20:01:10.967420 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r6j4l" event={"ID":"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec","Type":"ContainerDied","Data":"7372ee4e45a5f3cce7e2ff786d7314531acc1f0d8351fb1e1582def7a5cbc28a"} Dec 01 20:01:11 crc kubenswrapper[4960]: I1201 20:01:11.980696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"d0d9dc6db695566505c6d63af2eba5441e10ec489b203065c4bd2df26aa74c32"} Dec 01 20:01:11 crc kubenswrapper[4960]: I1201 20:01:11.980962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"5e36ab85d99a784814c8b962235dd80aafa3fe4594ebab991a431b1b3aefc870"} Dec 01 20:01:11 crc kubenswrapper[4960]: I1201 20:01:11.980973 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"41ab2d413967c064179bd9876245647b23cd7735b54eecd0775d875f3fbc2b41"} Dec 01 20:01:11 crc kubenswrapper[4960]: I1201 20:01:11.980981 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"ab67acef853f3f6167f363eb9024c76d3a0e6b52800730dc78427b506da3e51b"} Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.439354 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r6j4l" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.543481 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle\") pod \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.543543 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlk9g\" (UniqueName: \"kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g\") pod \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.543682 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data\") pod \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.543734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data\") pod \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\" (UID: \"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec\") " Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.552899 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" (UID: "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.553287 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g" (OuterVolumeSpecName: "kube-api-access-nlk9g") pod "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" (UID: "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec"). InnerVolumeSpecName "kube-api-access-nlk9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.581315 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" (UID: "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.599667 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data" (OuterVolumeSpecName: "config-data") pod "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" (UID: "ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.646489 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.647017 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlk9g\" (UniqueName: \"kubernetes.io/projected/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-kube-api-access-nlk9g\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.647188 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:12 crc kubenswrapper[4960]: I1201 20:01:12.647270 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.011310 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"9a49a37feee59225d4cef84b223c4159432d033dd58b6cfc04bbbbe3a9011feb"} Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.011372 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"69d162075894765fe69327a613911275748782da031e5cd64ca0e803966a32bd"} Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.011395 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"725698af-f23c-4262-96a4-2bfff12018fa","Type":"ContainerStarted","Data":"61d5ad705532a8c94ad1c2c998c9c1df711453c14491e2235aa703c348f7b182"} Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.017139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r6j4l" event={"ID":"ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec","Type":"ContainerDied","Data":"32503d58ab175a217b9cb1e5ad97e07cdeb062eaa14a56b4d4663e9309ffa78b"} Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.017210 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32503d58ab175a217b9cb1e5ad97e07cdeb062eaa14a56b4d4663e9309ffa78b" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.017312 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r6j4l" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.077452 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.715824884 podStartE2EDuration="44.077425147s" podCreationTimestamp="2025-12-01 20:00:29 +0000 UTC" firstStartedPulling="2025-12-01 20:01:03.71401555 +0000 UTC m=+1299.001507219" lastFinishedPulling="2025-12-01 20:01:11.075615813 +0000 UTC m=+1306.363107482" observedRunningTime="2025-12-01 20:01:13.073510296 +0000 UTC m=+1308.361002005" watchObservedRunningTime="2025-12-01 20:01:13.077425147 +0000 UTC m=+1308.364916866" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.407229 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-p79h2"] Dec 01 20:01:13 crc kubenswrapper[4960]: E1201 20:01:13.407728 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" containerName="glance-db-sync" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.407743 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" containerName="glance-db-sync" Dec 01 20:01:13 crc kubenswrapper[4960]: E1201 20:01:13.407757 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eea7d45-e33e-4bec-9810-bee3cff9c612" containerName="ovn-config" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.407763 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eea7d45-e33e-4bec-9810-bee3cff9c612" containerName="ovn-config" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.407951 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eea7d45-e33e-4bec-9810-bee3cff9c612" containerName="ovn-config" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.407978 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" containerName="glance-db-sync" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.409062 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.424270 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-p79h2"] Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.491858 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-p79h2"] Dec 01 20:01:13 crc kubenswrapper[4960]: E1201 20:01:13.492523 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-vv9fc ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" podUID="6ece0e6e-3a50-422b-83d3-aa63a2afdd99" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.516228 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.527835 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.528920 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.536381 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.570507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.570866 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.570896 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.571008 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.571047 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv9fc\" (UniqueName: \"kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672775 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672838 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672864 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672922 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672942 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv9fc\" (UniqueName: \"kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.672964 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxrjx\" (UniqueName: \"kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.673232 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.673309 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.673342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.673393 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.673423 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.674781 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.674795 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.675027 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.675060 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.692235 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv9fc\" (UniqueName: \"kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc\") pod \"dnsmasq-dns-5b946c75cc-p79h2\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.774838 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.774881 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.774917 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.774941 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxrjx\" (UniqueName: \"kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.775006 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.775022 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.775990 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.776108 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.776350 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.776441 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.776863 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.792389 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxrjx\" (UniqueName: \"kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx\") pod \"dnsmasq-dns-74f6bcbc87-t42dw\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:13 crc kubenswrapper[4960]: I1201 20:01:13.846224 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.028147 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.044061 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.182964 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv9fc\" (UniqueName: \"kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc\") pod \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.183410 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb\") pod \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.183616 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc\") pod \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.183666 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb\") pod \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.183695 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config\") pod \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\" (UID: \"6ece0e6e-3a50-422b-83d3-aa63a2afdd99\") " Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.185023 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ece0e6e-3a50-422b-83d3-aa63a2afdd99" (UID: "6ece0e6e-3a50-422b-83d3-aa63a2afdd99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.185154 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config" (OuterVolumeSpecName: "config") pod "6ece0e6e-3a50-422b-83d3-aa63a2afdd99" (UID: "6ece0e6e-3a50-422b-83d3-aa63a2afdd99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.185342 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ece0e6e-3a50-422b-83d3-aa63a2afdd99" (UID: "6ece0e6e-3a50-422b-83d3-aa63a2afdd99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.189546 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ece0e6e-3a50-422b-83d3-aa63a2afdd99" (UID: "6ece0e6e-3a50-422b-83d3-aa63a2afdd99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.193372 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc" (OuterVolumeSpecName: "kube-api-access-vv9fc") pod "6ece0e6e-3a50-422b-83d3-aa63a2afdd99" (UID: "6ece0e6e-3a50-422b-83d3-aa63a2afdd99"). InnerVolumeSpecName "kube-api-access-vv9fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.286027 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.286060 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.286071 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.286080 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv9fc\" (UniqueName: \"kubernetes.io/projected/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-kube-api-access-vv9fc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.286090 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ece0e6e-3a50-422b-83d3-aa63a2afdd99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.337575 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:14 crc kubenswrapper[4960]: W1201 20:01:14.340616 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod960cddd2_52b0_4c62_b45f_fca88b7b6183.slice/crio-b971ca8bb725f3535853ac9ed38d912e6dc5bef00baa75fe105bfea7a6b17537 WatchSource:0}: Error finding container b971ca8bb725f3535853ac9ed38d912e6dc5bef00baa75fe105bfea7a6b17537: Status 404 returned error can't find the container with id b971ca8bb725f3535853ac9ed38d912e6dc5bef00baa75fe105bfea7a6b17537 Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.357511 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.470413 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.704801 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-z7sq7"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.709820 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.724511 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z7sq7"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.795794 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.796186 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw4z2\" (UniqueName: \"kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.803897 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qb9hb"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.806034 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.813588 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qb9hb"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.898291 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz8l8\" (UniqueName: \"kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.898381 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.898465 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw4z2\" (UniqueName: \"kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.898531 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.899405 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.907137 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-9a0f-account-create-update-brw2l"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.909012 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.912481 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.926625 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw4z2\" (UniqueName: \"kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2\") pod \"cinder-db-create-z7sq7\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.926883 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9a0f-account-create-update-brw2l"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.998960 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-create-fbxww"] Dec 01 20:01:14 crc kubenswrapper[4960]: I1201 20:01:14.999937 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.000146 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsz8v\" (UniqueName: \"kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.000184 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz8l8\" (UniqueName: \"kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.000212 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.000255 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.000675 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.008984 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8eff-account-create-update-2p9kt"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.010359 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.011912 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.023545 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-fbxww"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.032093 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz8l8\" (UniqueName: \"kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8\") pod \"barbican-db-create-qb9hb\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.042170 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.067614 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8eff-account-create-update-2p9kt"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.073894 4960 generic.go:334] "Generic (PLEG): container finished" podID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerID="c7af915a692e7cca853ac5179a0bce304581559f2b3855d014b11404e85d792d" exitCode=0 Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.073953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" event={"ID":"960cddd2-52b0-4c62-b45f-fca88b7b6183","Type":"ContainerDied","Data":"c7af915a692e7cca853ac5179a0bce304581559f2b3855d014b11404e85d792d"} Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.073979 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" event={"ID":"960cddd2-52b0-4c62-b45f-fca88b7b6183","Type":"ContainerStarted","Data":"b971ca8bb725f3535853ac9ed38d912e6dc5bef00baa75fe105bfea7a6b17537"} Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.086589 4960 generic.go:334] "Generic (PLEG): container finished" podID="9a4f20da-f7ef-4710-9222-77baad783676" containerID="3bb62657767591097ba8f70da2c3f0e435069c20d5903734eaf578dbca202f99" exitCode=0 Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.086668 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-p79h2" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.087429 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerDied","Data":"3bb62657767591097ba8f70da2c3f0e435069c20d5903734eaf578dbca202f99"} Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.107724 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.107787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dblv8\" (UniqueName: \"kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.107826 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.107927 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsz8v\" (UniqueName: \"kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.107972 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.108098 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crwqp\" (UniqueName: \"kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.112057 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.124385 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.178895 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsz8v\" (UniqueName: \"kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v\") pod \"barbican-9a0f-account-create-update-brw2l\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.201903 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-fl9dv"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.203297 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.208209 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mm9q9" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.208309 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.208432 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.208604 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.209486 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.209542 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dblv8\" (UniqueName: \"kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.209568 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.209795 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crwqp\" (UniqueName: \"kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.210774 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.210795 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.231181 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-fl9dv"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.242342 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dblv8\" (UniqueName: \"kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8\") pod \"cloudkitty-db-create-fbxww\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.247233 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crwqp\" (UniqueName: \"kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp\") pod \"cinder-8eff-account-create-update-2p9kt\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.300155 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.310486 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-87eb-account-create-update-xw6fw"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.311303 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8fwc\" (UniqueName: \"kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.311426 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.311560 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.312939 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.318829 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.326314 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.332461 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-db-secret" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.364886 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-87eb-account-create-update-xw6fw"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.367909 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-p79h2"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.379682 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-p79h2"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.413794 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.413853 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.413896 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8fwc\" (UniqueName: \"kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.413988 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.414012 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9trh\" (UniqueName: \"kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.421403 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-cc43-account-create-update-kkd2v"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.422502 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.422901 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.430660 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.439200 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.441078 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc43-account-create-update-kkd2v"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.471936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8fwc\" (UniqueName: \"kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc\") pod \"keystone-db-sync-fl9dv\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.505239 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-frxcr"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.506622 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.516216 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzhr4\" (UniqueName: \"kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.516343 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.516397 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.516444 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9trh\" (UniqueName: \"kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.517538 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.522123 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-frxcr"] Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.529243 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.578639 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9trh\" (UniqueName: \"kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh\") pod \"cloudkitty-87eb-account-create-update-xw6fw\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.617860 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.617940 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnn5w\" (UniqueName: \"kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.617979 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzhr4\" (UniqueName: \"kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.618044 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.619263 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.670854 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzhr4\" (UniqueName: \"kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4\") pod \"neutron-cc43-account-create-update-kkd2v\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.722365 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnn5w\" (UniqueName: \"kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.722496 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.723451 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.729964 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.753184 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.805501 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnn5w\" (UniqueName: \"kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w\") pod \"neutron-db-create-frxcr\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:15 crc kubenswrapper[4960]: I1201 20:01:15.899008 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z7sq7"] Dec 01 20:01:16 crc kubenswrapper[4960]: W1201 20:01:16.071677 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe342ccd_a32e_42ba_9018_c4fe438c9d8c.slice/crio-f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f WatchSource:0}: Error finding container f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f: Status 404 returned error can't find the container with id f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.072100 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.073679 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qb9hb"] Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.129359 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" event={"ID":"960cddd2-52b0-4c62-b45f-fca88b7b6183","Type":"ContainerStarted","Data":"b4bfe30915785ce3ecbdb5707d2f4bd7bd92a15c06e4f8a7e4fb4fa71b8992e5"} Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.129801 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.136846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerStarted","Data":"172286349de35a34001d18653664481f6ce410802d801fe81c49945420d8e331"} Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.140057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z7sq7" event={"ID":"923a3511-6d9b-4143-867b-552769cdbd94","Type":"ContainerStarted","Data":"78af23707ea306bb0c2a214f86b6ff65a2716f09236a30708f5db4905f46a418"} Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.144501 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qb9hb" event={"ID":"fe342ccd-a32e-42ba-9018-c4fe438c9d8c","Type":"ContainerStarted","Data":"f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f"} Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.191334 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" podStartSLOduration=3.191316424 podStartE2EDuration="3.191316424s" podCreationTimestamp="2025-12-01 20:01:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:16.15752894 +0000 UTC m=+1311.445020609" watchObservedRunningTime="2025-12-01 20:01:16.191316424 +0000 UTC m=+1311.478808093" Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.202125 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9a0f-account-create-update-brw2l"] Dec 01 20:01:16 crc kubenswrapper[4960]: W1201 20:01:16.256050 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podead81fb5_9e98_4b2a_8e02_8d9b82bef13b.slice/crio-2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7 WatchSource:0}: Error finding container 2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7: Status 404 returned error can't find the container with id 2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7 Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.261139 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8eff-account-create-update-2p9kt"] Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.366507 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-create-fbxww"] Dec 01 20:01:16 crc kubenswrapper[4960]: W1201 20:01:16.369739 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2ce664d_4e8c_479a_95af_9f38cdeb98eb.slice/crio-ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907 WatchSource:0}: Error finding container ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907: Status 404 returned error can't find the container with id ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907 Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.560650 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-fl9dv"] Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.706437 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc43-account-create-update-kkd2v"] Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.715781 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-87eb-account-create-update-xw6fw"] Dec 01 20:01:16 crc kubenswrapper[4960]: W1201 20:01:16.718357 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0190c0ae_497c_404f_b7f0_a57250e0339e.slice/crio-f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750 WatchSource:0}: Error finding container f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750: Status 404 returned error can't find the container with id f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750 Dec 01 20:01:16 crc kubenswrapper[4960]: I1201 20:01:16.823869 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-frxcr"] Dec 01 20:01:16 crc kubenswrapper[4960]: W1201 20:01:16.835474 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd132ae6_3b94_42f6_804a_29e58ba2d671.slice/crio-4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1 WatchSource:0}: Error finding container 4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1: Status 404 returned error can't find the container with id 4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.156471 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" event={"ID":"8b811ebc-3885-4b30-8e97-373aaaff228e","Type":"ContainerStarted","Data":"00d0dbf5e8d3f15432f74b81e5abd5cd84e4db185c5eece1c6ec9a7762218bf2"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.156519 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" event={"ID":"8b811ebc-3885-4b30-8e97-373aaaff228e","Type":"ContainerStarted","Data":"784ac1e37e27812b6065e315a2cfc04cb336ced7b3901e0da90effd11c6ec493"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.159449 4960 generic.go:334] "Generic (PLEG): container finished" podID="ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" containerID="e1eb5f725af2d68550a6c50a198ec70012389e94e0f620ce9871be973a497d5c" exitCode=0 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.159524 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9a0f-account-create-update-brw2l" event={"ID":"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b","Type":"ContainerDied","Data":"e1eb5f725af2d68550a6c50a198ec70012389e94e0f620ce9871be973a497d5c"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.159545 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9a0f-account-create-update-brw2l" event={"ID":"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b","Type":"ContainerStarted","Data":"2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.161069 4960 generic.go:334] "Generic (PLEG): container finished" podID="fe342ccd-a32e-42ba-9018-c4fe438c9d8c" containerID="f3123874e2573b16f320da0a9f0954e5a50cb3c67a9f6fff126bf8cabdf4a58f" exitCode=0 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.161109 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qb9hb" event={"ID":"fe342ccd-a32e-42ba-9018-c4fe438c9d8c","Type":"ContainerDied","Data":"f3123874e2573b16f320da0a9f0954e5a50cb3c67a9f6fff126bf8cabdf4a58f"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.162422 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fl9dv" event={"ID":"948143c4-753e-4129-9712-825a69981406","Type":"ContainerStarted","Data":"28c667336b95c9bce3ea942aed28d7a9578810739388065f98b4636c33cde0a4"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.164055 4960 generic.go:334] "Generic (PLEG): container finished" podID="c2ce664d-4e8c-479a-95af-9f38cdeb98eb" containerID="d74c66b34068cfe52a3d451125cf9118482566ac828a4f6648ca6c19092f38ea" exitCode=0 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.164168 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-fbxww" event={"ID":"c2ce664d-4e8c-479a-95af-9f38cdeb98eb","Type":"ContainerDied","Data":"d74c66b34068cfe52a3d451125cf9118482566ac828a4f6648ca6c19092f38ea"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.164217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-fbxww" event={"ID":"c2ce664d-4e8c-479a-95af-9f38cdeb98eb","Type":"ContainerStarted","Data":"ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.165727 4960 generic.go:334] "Generic (PLEG): container finished" podID="923a3511-6d9b-4143-867b-552769cdbd94" containerID="effc3da163f3ae4d02dc19621bdb0b5f2fc63641015465208e6e6ac287919008" exitCode=0 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.165773 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z7sq7" event={"ID":"923a3511-6d9b-4143-867b-552769cdbd94","Type":"ContainerDied","Data":"effc3da163f3ae4d02dc19621bdb0b5f2fc63641015465208e6e6ac287919008"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.167220 4960 generic.go:334] "Generic (PLEG): container finished" podID="b513d8e1-b2fa-476d-b925-12ce32db7390" containerID="25522c134c0c582eb94b175a5405841ade1504a5a3af093d7c66c43f52b6a170" exitCode=0 Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.167288 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8eff-account-create-update-2p9kt" event={"ID":"b513d8e1-b2fa-476d-b925-12ce32db7390","Type":"ContainerDied","Data":"25522c134c0c582eb94b175a5405841ade1504a5a3af093d7c66c43f52b6a170"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.167303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8eff-account-create-update-2p9kt" event={"ID":"b513d8e1-b2fa-476d-b925-12ce32db7390","Type":"ContainerStarted","Data":"48f87890feb042d73765aa0b9093e55a075f6287c6800f86f0dfe7a71af7808c"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.169318 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frxcr" event={"ID":"bd132ae6-3b94-42f6-804a-29e58ba2d671","Type":"ContainerStarted","Data":"9c6fcea4c2b54ae1ac37450f075a146cbdbd2bc74ac325c78ed8a2707387c55c"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.169377 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frxcr" event={"ID":"bd132ae6-3b94-42f6-804a-29e58ba2d671","Type":"ContainerStarted","Data":"4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.172275 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc43-account-create-update-kkd2v" event={"ID":"0190c0ae-497c-404f-b7f0-a57250e0339e","Type":"ContainerStarted","Data":"2b69b3a9445b35adc8c89447ecebdb94e10356d2e46bb006b364cc3ef2f92609"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.172319 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc43-account-create-update-kkd2v" event={"ID":"0190c0ae-497c-404f-b7f0-a57250e0339e","Type":"ContainerStarted","Data":"f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750"} Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.182958 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" podStartSLOduration=2.182937219 podStartE2EDuration="2.182937219s" podCreationTimestamp="2025-12-01 20:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:17.178031368 +0000 UTC m=+1312.465523047" watchObservedRunningTime="2025-12-01 20:01:17.182937219 +0000 UTC m=+1312.470428888" Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.213943 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-cc43-account-create-update-kkd2v" podStartSLOduration=2.213921239 podStartE2EDuration="2.213921239s" podCreationTimestamp="2025-12-01 20:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:17.208956415 +0000 UTC m=+1312.496448084" watchObservedRunningTime="2025-12-01 20:01:17.213921239 +0000 UTC m=+1312.501412908" Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.246962 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-frxcr" podStartSLOduration=2.24693305 podStartE2EDuration="2.24693305s" podCreationTimestamp="2025-12-01 20:01:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:17.242032609 +0000 UTC m=+1312.529524278" watchObservedRunningTime="2025-12-01 20:01:17.24693305 +0000 UTC m=+1312.534424719" Dec 01 20:01:17 crc kubenswrapper[4960]: I1201 20:01:17.334718 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ece0e6e-3a50-422b-83d3-aa63a2afdd99" path="/var/lib/kubelet/pods/6ece0e6e-3a50-422b-83d3-aa63a2afdd99/volumes" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.182232 4960 generic.go:334] "Generic (PLEG): container finished" podID="bd132ae6-3b94-42f6-804a-29e58ba2d671" containerID="9c6fcea4c2b54ae1ac37450f075a146cbdbd2bc74ac325c78ed8a2707387c55c" exitCode=0 Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.182336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frxcr" event={"ID":"bd132ae6-3b94-42f6-804a-29e58ba2d671","Type":"ContainerDied","Data":"9c6fcea4c2b54ae1ac37450f075a146cbdbd2bc74ac325c78ed8a2707387c55c"} Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.185027 4960 generic.go:334] "Generic (PLEG): container finished" podID="0190c0ae-497c-404f-b7f0-a57250e0339e" containerID="2b69b3a9445b35adc8c89447ecebdb94e10356d2e46bb006b364cc3ef2f92609" exitCode=0 Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.185073 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc43-account-create-update-kkd2v" event={"ID":"0190c0ae-497c-404f-b7f0-a57250e0339e","Type":"ContainerDied","Data":"2b69b3a9445b35adc8c89447ecebdb94e10356d2e46bb006b364cc3ef2f92609"} Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.187198 4960 generic.go:334] "Generic (PLEG): container finished" podID="8b811ebc-3885-4b30-8e97-373aaaff228e" containerID="00d0dbf5e8d3f15432f74b81e5abd5cd84e4db185c5eece1c6ec9a7762218bf2" exitCode=0 Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.187243 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" event={"ID":"8b811ebc-3885-4b30-8e97-373aaaff228e","Type":"ContainerDied","Data":"00d0dbf5e8d3f15432f74b81e5abd5cd84e4db185c5eece1c6ec9a7762218bf2"} Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.668739 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.810663 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crwqp\" (UniqueName: \"kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp\") pod \"b513d8e1-b2fa-476d-b925-12ce32db7390\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.810729 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts\") pod \"b513d8e1-b2fa-476d-b925-12ce32db7390\" (UID: \"b513d8e1-b2fa-476d-b925-12ce32db7390\") " Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.812935 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b513d8e1-b2fa-476d-b925-12ce32db7390" (UID: "b513d8e1-b2fa-476d-b925-12ce32db7390"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.818526 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp" (OuterVolumeSpecName: "kube-api-access-crwqp") pod "b513d8e1-b2fa-476d-b925-12ce32db7390" (UID: "b513d8e1-b2fa-476d-b925-12ce32db7390"). InnerVolumeSpecName "kube-api-access-crwqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.898285 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.904333 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.914427 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crwqp\" (UniqueName: \"kubernetes.io/projected/b513d8e1-b2fa-476d-b925-12ce32db7390-kube-api-access-crwqp\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.914457 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b513d8e1-b2fa-476d-b925-12ce32db7390-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.916655 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:18 crc kubenswrapper[4960]: I1201 20:01:18.935910 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016328 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsz8v\" (UniqueName: \"kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v\") pod \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016409 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dblv8\" (UniqueName: \"kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8\") pod \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016497 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts\") pod \"923a3511-6d9b-4143-867b-552769cdbd94\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016675 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts\") pod \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\" (UID: \"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016743 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz8l8\" (UniqueName: \"kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8\") pod \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016782 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts\") pod \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\" (UID: \"fe342ccd-a32e-42ba-9018-c4fe438c9d8c\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.016922 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw4z2\" (UniqueName: \"kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2\") pod \"923a3511-6d9b-4143-867b-552769cdbd94\" (UID: \"923a3511-6d9b-4143-867b-552769cdbd94\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.017029 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts\") pod \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\" (UID: \"c2ce664d-4e8c-479a-95af-9f38cdeb98eb\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.017332 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "923a3511-6d9b-4143-867b-552769cdbd94" (UID: "923a3511-6d9b-4143-867b-552769cdbd94"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.017351 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" (UID: "ead81fb5-9e98-4b2a-8e02-8d9b82bef13b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.017848 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/923a3511-6d9b-4143-867b-552769cdbd94-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.017872 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.018045 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe342ccd-a32e-42ba-9018-c4fe438c9d8c" (UID: "fe342ccd-a32e-42ba-9018-c4fe438c9d8c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.018047 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2ce664d-4e8c-479a-95af-9f38cdeb98eb" (UID: "c2ce664d-4e8c-479a-95af-9f38cdeb98eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.020594 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2" (OuterVolumeSpecName: "kube-api-access-kw4z2") pod "923a3511-6d9b-4143-867b-552769cdbd94" (UID: "923a3511-6d9b-4143-867b-552769cdbd94"). InnerVolumeSpecName "kube-api-access-kw4z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.021083 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v" (OuterVolumeSpecName: "kube-api-access-xsz8v") pod "ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" (UID: "ead81fb5-9e98-4b2a-8e02-8d9b82bef13b"). InnerVolumeSpecName "kube-api-access-xsz8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.022198 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8" (OuterVolumeSpecName: "kube-api-access-dblv8") pod "c2ce664d-4e8c-479a-95af-9f38cdeb98eb" (UID: "c2ce664d-4e8c-479a-95af-9f38cdeb98eb"). InnerVolumeSpecName "kube-api-access-dblv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.022838 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8" (OuterVolumeSpecName: "kube-api-access-cz8l8") pod "fe342ccd-a32e-42ba-9018-c4fe438c9d8c" (UID: "fe342ccd-a32e-42ba-9018-c4fe438c9d8c"). InnerVolumeSpecName "kube-api-access-cz8l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119606 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsz8v\" (UniqueName: \"kubernetes.io/projected/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b-kube-api-access-xsz8v\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119637 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dblv8\" (UniqueName: \"kubernetes.io/projected/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-kube-api-access-dblv8\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119648 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz8l8\" (UniqueName: \"kubernetes.io/projected/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-kube-api-access-cz8l8\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119657 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe342ccd-a32e-42ba-9018-c4fe438c9d8c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119665 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw4z2\" (UniqueName: \"kubernetes.io/projected/923a3511-6d9b-4143-867b-552769cdbd94-kube-api-access-kw4z2\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.119675 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2ce664d-4e8c-479a-95af-9f38cdeb98eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.198839 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z7sq7" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.198832 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z7sq7" event={"ID":"923a3511-6d9b-4143-867b-552769cdbd94","Type":"ContainerDied","Data":"78af23707ea306bb0c2a214f86b6ff65a2716f09236a30708f5db4905f46a418"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.198964 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78af23707ea306bb0c2a214f86b6ff65a2716f09236a30708f5db4905f46a418" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.201806 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9a0f-account-create-update-brw2l" event={"ID":"ead81fb5-9e98-4b2a-8e02-8d9b82bef13b","Type":"ContainerDied","Data":"2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.201830 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9a0f-account-create-update-brw2l" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.201843 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2be11d0c38f02ee0a9556dbb26000bbc9530cc74076847beacdf9337e7d996d7" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.204035 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qb9hb" event={"ID":"fe342ccd-a32e-42ba-9018-c4fe438c9d8c","Type":"ContainerDied","Data":"f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.204061 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5d03f331952ffabfc8fb9bf7054db69de0be7f4024d8db59c1034e54b3ce54f" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.204102 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qb9hb" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.211229 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerStarted","Data":"65fd62276c4ae519874c7ce6f08fb0422b926aad0601a70035d674b03d3b3d96"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.211279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9a4f20da-f7ef-4710-9222-77baad783676","Type":"ContainerStarted","Data":"5826146d903d911573e0934f496a14cd969bd104f48684e1e1eb12ef62a9a890"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.215009 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8eff-account-create-update-2p9kt" event={"ID":"b513d8e1-b2fa-476d-b925-12ce32db7390","Type":"ContainerDied","Data":"48f87890feb042d73765aa0b9093e55a075f6287c6800f86f0dfe7a71af7808c"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.215046 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48f87890feb042d73765aa0b9093e55a075f6287c6800f86f0dfe7a71af7808c" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.215094 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8eff-account-create-update-2p9kt" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.222577 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-create-fbxww" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.222631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-create-fbxww" event={"ID":"c2ce664d-4e8c-479a-95af-9f38cdeb98eb","Type":"ContainerDied","Data":"ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907"} Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.222691 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad08400c111c5d16bb2ea574c0cb28159fb4d18d2f4849ecc4589731cf2d7907" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.247049 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.247029352 podStartE2EDuration="16.247029352s" podCreationTimestamp="2025-12-01 20:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:19.243624856 +0000 UTC m=+1314.531116545" watchObservedRunningTime="2025-12-01 20:01:19.247029352 +0000 UTC m=+1314.534521021" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.348004 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.348718 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.355490 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.792719 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.799319 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.809092 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.830991 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts\") pod \"bd132ae6-3b94-42f6-804a-29e58ba2d671\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.831194 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnn5w\" (UniqueName: \"kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w\") pod \"bd132ae6-3b94-42f6-804a-29e58ba2d671\" (UID: \"bd132ae6-3b94-42f6-804a-29e58ba2d671\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.831678 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd132ae6-3b94-42f6-804a-29e58ba2d671" (UID: "bd132ae6-3b94-42f6-804a-29e58ba2d671"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.842320 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w" (OuterVolumeSpecName: "kube-api-access-jnn5w") pod "bd132ae6-3b94-42f6-804a-29e58ba2d671" (UID: "bd132ae6-3b94-42f6-804a-29e58ba2d671"). InnerVolumeSpecName "kube-api-access-jnn5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.936541 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzhr4\" (UniqueName: \"kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4\") pod \"0190c0ae-497c-404f-b7f0-a57250e0339e\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.936659 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts\") pod \"0190c0ae-497c-404f-b7f0-a57250e0339e\" (UID: \"0190c0ae-497c-404f-b7f0-a57250e0339e\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.936835 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts\") pod \"8b811ebc-3885-4b30-8e97-373aaaff228e\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.936878 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9trh\" (UniqueName: \"kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh\") pod \"8b811ebc-3885-4b30-8e97-373aaaff228e\" (UID: \"8b811ebc-3885-4b30-8e97-373aaaff228e\") " Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.937609 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0190c0ae-497c-404f-b7f0-a57250e0339e" (UID: "0190c0ae-497c-404f-b7f0-a57250e0339e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.937659 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd132ae6-3b94-42f6-804a-29e58ba2d671-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.937678 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnn5w\" (UniqueName: \"kubernetes.io/projected/bd132ae6-3b94-42f6-804a-29e58ba2d671-kube-api-access-jnn5w\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.938317 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b811ebc-3885-4b30-8e97-373aaaff228e" (UID: "8b811ebc-3885-4b30-8e97-373aaaff228e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.940568 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh" (OuterVolumeSpecName: "kube-api-access-h9trh") pod "8b811ebc-3885-4b30-8e97-373aaaff228e" (UID: "8b811ebc-3885-4b30-8e97-373aaaff228e"). InnerVolumeSpecName "kube-api-access-h9trh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:19 crc kubenswrapper[4960]: I1201 20:01:19.940666 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4" (OuterVolumeSpecName: "kube-api-access-jzhr4") pod "0190c0ae-497c-404f-b7f0-a57250e0339e" (UID: "0190c0ae-497c-404f-b7f0-a57250e0339e"). InnerVolumeSpecName "kube-api-access-jzhr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.039023 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzhr4\" (UniqueName: \"kubernetes.io/projected/0190c0ae-497c-404f-b7f0-a57250e0339e-kube-api-access-jzhr4\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.039065 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0190c0ae-497c-404f-b7f0-a57250e0339e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.039075 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b811ebc-3885-4b30-8e97-373aaaff228e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.039084 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9trh\" (UniqueName: \"kubernetes.io/projected/8b811ebc-3885-4b30-8e97-373aaaff228e-kube-api-access-h9trh\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.237801 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frxcr" event={"ID":"bd132ae6-3b94-42f6-804a-29e58ba2d671","Type":"ContainerDied","Data":"4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1"} Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.237860 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d176dd9e76d862ecb00a20e1fe861adee944feb879795e56efc52e1e8e34ae1" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.237924 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frxcr" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.240711 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc43-account-create-update-kkd2v" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.241315 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc43-account-create-update-kkd2v" event={"ID":"0190c0ae-497c-404f-b7f0-a57250e0339e","Type":"ContainerDied","Data":"f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750"} Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.241360 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f48f5bb56bb82dd59c1050f028b54f298152d56c4570751927dfe81fa1350750" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.243458 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" event={"ID":"8b811ebc-3885-4b30-8e97-373aaaff228e","Type":"ContainerDied","Data":"784ac1e37e27812b6065e315a2cfc04cb336ced7b3901e0da90effd11c6ec493"} Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.243490 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="784ac1e37e27812b6065e315a2cfc04cb336ced7b3901e0da90effd11c6ec493" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.243533 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-87eb-account-create-update-xw6fw" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.250799 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 01 20:01:20 crc kubenswrapper[4960]: I1201 20:01:20.272008 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-lokistack-ingester-0" Dec 01 20:01:23 crc kubenswrapper[4960]: I1201 20:01:23.848381 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:23 crc kubenswrapper[4960]: I1201 20:01:23.942298 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:01:23 crc kubenswrapper[4960]: I1201 20:01:23.942584 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-p7pmg" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="dnsmasq-dns" containerID="cri-o://4abbc71a786a06dda888cbd461b9fa2d992debe605a22a578320c0d407cb1ddb" gracePeriod=10 Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.296212 4960 generic.go:334] "Generic (PLEG): container finished" podID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerID="4abbc71a786a06dda888cbd461b9fa2d992debe605a22a578320c0d407cb1ddb" exitCode=0 Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.296297 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p7pmg" event={"ID":"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d","Type":"ContainerDied","Data":"4abbc71a786a06dda888cbd461b9fa2d992debe605a22a578320c0d407cb1ddb"} Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.303424 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fl9dv" event={"ID":"948143c4-753e-4129-9712-825a69981406","Type":"ContainerStarted","Data":"87de426e42aba84c08c10a5023f81eb1b83297cfb0c8ae187c8fdd19c38a2b2d"} Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.332780 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-fl9dv" podStartSLOduration=2.722094898 podStartE2EDuration="9.332757835s" podCreationTimestamp="2025-12-01 20:01:15 +0000 UTC" firstStartedPulling="2025-12-01 20:01:16.575414794 +0000 UTC m=+1311.862906463" lastFinishedPulling="2025-12-01 20:01:23.186077731 +0000 UTC m=+1318.473569400" observedRunningTime="2025-12-01 20:01:24.324669805 +0000 UTC m=+1319.612161474" watchObservedRunningTime="2025-12-01 20:01:24.332757835 +0000 UTC m=+1319.620249504" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.461753 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.487041 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb\") pod \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.487902 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config\") pod \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.487948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s58r\" (UniqueName: \"kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r\") pod \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.488033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb\") pod \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.488207 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc\") pod \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\" (UID: \"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d\") " Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.493681 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r" (OuterVolumeSpecName: "kube-api-access-4s58r") pod "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" (UID: "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d"). InnerVolumeSpecName "kube-api-access-4s58r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.538276 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config" (OuterVolumeSpecName: "config") pod "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" (UID: "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.546951 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" (UID: "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.558235 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" (UID: "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.561319 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" (UID: "febe2257-7e6e-484a-9ea5-7bea7dbc0a1d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.591262 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.591298 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.591307 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.591318 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:24 crc kubenswrapper[4960]: I1201 20:01:24.591330 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s58r\" (UniqueName: \"kubernetes.io/projected/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d-kube-api-access-4s58r\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.314971 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-p7pmg" event={"ID":"febe2257-7e6e-484a-9ea5-7bea7dbc0a1d","Type":"ContainerDied","Data":"a52d94d6ccd24ac7f0b3db47a152423a89224b66104078690d5c8c2fa2195a8b"} Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.315306 4960 scope.go:117] "RemoveContainer" containerID="4abbc71a786a06dda888cbd461b9fa2d992debe605a22a578320c0d407cb1ddb" Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.315196 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-p7pmg" Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.351636 4960 scope.go:117] "RemoveContainer" containerID="1d403eec1c842f6b6acf20822bb58f7e2e682111b6d21a117d9d496e6dde017a" Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.375309 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:01:25 crc kubenswrapper[4960]: I1201 20:01:25.385194 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-p7pmg"] Dec 01 20:01:27 crc kubenswrapper[4960]: I1201 20:01:27.343666 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" path="/var/lib/kubelet/pods/febe2257-7e6e-484a-9ea5-7bea7dbc0a1d/volumes" Dec 01 20:01:27 crc kubenswrapper[4960]: I1201 20:01:27.346723 4960 generic.go:334] "Generic (PLEG): container finished" podID="948143c4-753e-4129-9712-825a69981406" containerID="87de426e42aba84c08c10a5023f81eb1b83297cfb0c8ae187c8fdd19c38a2b2d" exitCode=0 Dec 01 20:01:27 crc kubenswrapper[4960]: I1201 20:01:27.346776 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fl9dv" event={"ID":"948143c4-753e-4129-9712-825a69981406","Type":"ContainerDied","Data":"87de426e42aba84c08c10a5023f81eb1b83297cfb0c8ae187c8fdd19c38a2b2d"} Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.696098 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.791592 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data\") pod \"948143c4-753e-4129-9712-825a69981406\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.791671 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle\") pod \"948143c4-753e-4129-9712-825a69981406\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.791797 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8fwc\" (UniqueName: \"kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc\") pod \"948143c4-753e-4129-9712-825a69981406\" (UID: \"948143c4-753e-4129-9712-825a69981406\") " Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.803594 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc" (OuterVolumeSpecName: "kube-api-access-q8fwc") pod "948143c4-753e-4129-9712-825a69981406" (UID: "948143c4-753e-4129-9712-825a69981406"). InnerVolumeSpecName "kube-api-access-q8fwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.819780 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "948143c4-753e-4129-9712-825a69981406" (UID: "948143c4-753e-4129-9712-825a69981406"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.840764 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data" (OuterVolumeSpecName: "config-data") pod "948143c4-753e-4129-9712-825a69981406" (UID: "948143c4-753e-4129-9712-825a69981406"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.894740 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8fwc\" (UniqueName: \"kubernetes.io/projected/948143c4-753e-4129-9712-825a69981406-kube-api-access-q8fwc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.895063 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:28 crc kubenswrapper[4960]: I1201 20:01:28.895160 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948143c4-753e-4129-9712-825a69981406-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.404414 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-fl9dv" event={"ID":"948143c4-753e-4129-9712-825a69981406","Type":"ContainerDied","Data":"28c667336b95c9bce3ea942aed28d7a9578810739388065f98b4636c33cde0a4"} Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.404468 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c667336b95c9bce3ea942aed28d7a9578810739388065f98b4636c33cde0a4" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.404539 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-fl9dv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644064 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644802 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="init" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644819 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="init" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644829 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0190c0ae-497c-404f-b7f0-a57250e0339e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644835 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0190c0ae-497c-404f-b7f0-a57250e0339e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644849 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948143c4-753e-4129-9712-825a69981406" containerName="keystone-db-sync" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644856 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="948143c4-753e-4129-9712-825a69981406" containerName="keystone-db-sync" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644870 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd132ae6-3b94-42f6-804a-29e58ba2d671" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644876 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd132ae6-3b94-42f6-804a-29e58ba2d671" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644892 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923a3511-6d9b-4143-867b-552769cdbd94" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644899 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="923a3511-6d9b-4143-867b-552769cdbd94" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644907 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="dnsmasq-dns" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644913 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="dnsmasq-dns" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644924 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ce664d-4e8c-479a-95af-9f38cdeb98eb" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644929 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ce664d-4e8c-479a-95af-9f38cdeb98eb" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644941 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b513d8e1-b2fa-476d-b925-12ce32db7390" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644946 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b513d8e1-b2fa-476d-b925-12ce32db7390" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644965 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b811ebc-3885-4b30-8e97-373aaaff228e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644970 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b811ebc-3885-4b30-8e97-373aaaff228e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644977 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe342ccd-a32e-42ba-9018-c4fe438c9d8c" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644983 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe342ccd-a32e-42ba-9018-c4fe438c9d8c" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: E1201 20:01:29.644991 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.644997 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645222 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe342ccd-a32e-42ba-9018-c4fe438c9d8c" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645235 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645253 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="febe2257-7e6e-484a-9ea5-7bea7dbc0a1d" containerName="dnsmasq-dns" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645270 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b811ebc-3885-4b30-8e97-373aaaff228e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645280 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ce664d-4e8c-479a-95af-9f38cdeb98eb" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645292 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd132ae6-3b94-42f6-804a-29e58ba2d671" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645308 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="948143c4-753e-4129-9712-825a69981406" containerName="keystone-db-sync" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645319 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="923a3511-6d9b-4143-867b-552769cdbd94" containerName="mariadb-database-create" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645334 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b513d8e1-b2fa-476d-b925-12ce32db7390" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.645349 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0190c0ae-497c-404f-b7f0-a57250e0339e" containerName="mariadb-account-create-update" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.646861 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.657844 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.711722 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-gm7vf"] Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.716222 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.723603 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.723639 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.723603 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.723803 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.724898 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.724929 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.724975 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.725008 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.725028 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.725088 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9hxl\" (UniqueName: \"kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.729922 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mm9q9" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.737586 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gm7vf"] Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827761 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vhxr\" (UniqueName: \"kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827810 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827834 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827883 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827900 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827949 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.827969 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.828021 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9hxl\" (UniqueName: \"kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.828046 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.828086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.828912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.829434 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.829927 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.830434 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.830960 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.856385 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9hxl\" (UniqueName: \"kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl\") pod \"dnsmasq-dns-847c4cc679-l7rjv\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929499 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929552 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vhxr\" (UniqueName: \"kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929619 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929671 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.929747 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.941942 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.944390 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.947190 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.947586 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.947911 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.957412 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.960285 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.960845 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.961627 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.963931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vhxr\" (UniqueName: \"kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr\") pod \"keystone-bootstrap-gm7vf\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.973344 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:29 crc kubenswrapper[4960]: I1201 20:01:29.986780 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039681 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039724 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039777 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2272f\" (UniqueName: \"kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039794 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039823 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039850 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.039878 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.043289 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.043732 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.067172 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-nd25g"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.068489 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.073801 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-245x8" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.074002 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.093826 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nd25g"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.112309 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qmjlr"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.113904 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.123306 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.123854 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.124064 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-t4stk" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.144588 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qmjlr"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145557 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145671 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145746 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145836 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2272f\" (UniqueName: \"kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145908 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.145986 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfrpt\" (UniqueName: \"kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.146063 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.146160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.146246 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.146329 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.149069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.155985 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.156598 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.161345 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.162212 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.162522 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.185274 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.192255 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2272f\" (UniqueName: \"kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f\") pod \"ceilometer-0\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.204156 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-l6rmd"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.204950 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.205454 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.209802 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.210026 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kp5px" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.210242 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.244355 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-26xd2"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.246322 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.247758 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdpjl\" (UniqueName: \"kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.247848 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.247871 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.247941 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfrpt\" (UniqueName: \"kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.247981 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.248002 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.248025 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.248044 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.250186 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.250322 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jf8bn" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.250491 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.258464 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.261312 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.318014 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.319744 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfrpt\" (UniqueName: \"kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt\") pod \"barbican-db-sync-nd25g\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.341553 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-26xd2"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350080 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350181 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdpjl\" (UniqueName: \"kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350208 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350249 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350280 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350320 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350345 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350369 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350383 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350436 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkqm9\" (UniqueName: \"kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350497 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350535 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5knkj\" (UniqueName: \"kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350577 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350592 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.350617 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hshz\" (UniqueName: \"kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.352723 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.355022 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.364094 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.364349 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.374602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdpjl\" (UniqueName: \"kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl\") pod \"placement-db-sync-qmjlr\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.377199 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l6rmd"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.428673 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-zdrn9"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.429973 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.433333 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.433489 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.433527 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.437952 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ndz6g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.440831 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-zdrn9"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.444971 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454182 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454267 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454306 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454332 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454351 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454372 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkqm9\" (UniqueName: \"kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454664 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454681 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5knkj\" (UniqueName: \"kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454723 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454741 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454770 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hshz\" (UniqueName: \"kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454795 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.454825 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.455376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.456019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.457102 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.458497 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.459137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.498274 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.498791 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.498896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.499097 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.500273 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.500694 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nd25g" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.501253 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.503141 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.515371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkqm9\" (UniqueName: \"kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9\") pod \"cinder-db-sync-l6rmd\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.516396 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qmjlr" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.522818 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5knkj\" (UniqueName: \"kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj\") pod \"neutron-db-sync-26xd2\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.540624 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hshz\" (UniqueName: \"kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz\") pod \"dnsmasq-dns-785d8bcb8c-6p8jz\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.556774 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.557425 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.557622 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgg4s\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.558447 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.558703 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.563792 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.605220 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.633594 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.660445 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.660498 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.660543 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgg4s\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.660594 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.660676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.665572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.666310 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.668795 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.670370 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.690755 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgg4s\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s\") pod \"cloudkitty-db-sync-zdrn9\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.760815 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.813435 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.815224 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.821053 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.821286 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.822550 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mb8dw" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.822760 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.846067 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.858750 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.867793 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868095 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78v99\" (UniqueName: \"kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868238 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868351 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868442 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868555 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.868942 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.902640 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.905126 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.909390 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.909386 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.911887 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970711 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970773 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhrbk\" (UniqueName: \"kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970823 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970853 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970872 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.970888 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971101 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971189 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971222 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971247 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971268 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971298 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971318 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971335 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971355 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78v99\" (UniqueName: \"kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.971372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.973085 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.978446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.983780 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.983821 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91baf06d123dd91aed63649f36b06b6d82695580db5bb21a6dba903e67f7a922/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.984076 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.988462 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gm7vf"] Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.992019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:30 crc kubenswrapper[4960]: I1201 20:01:30.995061 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:30.999636 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:30.999921 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78v99\" (UniqueName: \"kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119603 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119667 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119764 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119799 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.119916 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhrbk\" (UniqueName: \"kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.120041 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.120062 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.120809 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.121025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.168592 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhrbk\" (UniqueName: \"kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.168928 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.168970 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/720c6baa92eb97844a8354a6d35963468f1db26ad29581ee9ba5e8c56ff873e5/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.169160 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.170055 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.170649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.184095 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.190419 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.230014 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.395432 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.435270 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" event={"ID":"40de8a90-fdbc-4a2a-974f-d3e7d266af31","Type":"ContainerStarted","Data":"24780b5cb870b2bf602327de889ec5325763f142706015f6336b6333d11afe6f"} Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.437303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gm7vf" event={"ID":"081b4a0b-4b47-4daa-a407-0e9076b71363","Type":"ContainerStarted","Data":"c8e6f4d14686b9e39bddd3314e323b354a13e3c0afacfe894ea8c64a1a2d609e"} Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.467350 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qmjlr"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.580979 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nd25g"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.602996 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.604335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.882017 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l6rmd"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.910185 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.918988 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-26xd2"] Dec 01 20:01:31 crc kubenswrapper[4960]: I1201 20:01:31.951370 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-zdrn9"] Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.209614 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:32 crc kubenswrapper[4960]: W1201 20:01:32.210596 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod036dc67e_619f_46e2_aec9_97c2ee5c74a7.slice/crio-9cef3201d3ad46d40d8fe97c1b98560e83a53c65cf8425300e8a9ffe432661f2 WatchSource:0}: Error finding container 9cef3201d3ad46d40d8fe97c1b98560e83a53c65cf8425300e8a9ffe432661f2: Status 404 returned error can't find the container with id 9cef3201d3ad46d40d8fe97c1b98560e83a53c65cf8425300e8a9ffe432661f2 Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.379254 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:32 crc kubenswrapper[4960]: W1201 20:01:32.384648 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04346d5b_21df_4897_a458_6b8cde19298d.slice/crio-2067c9beafc471fc2df3c09f452916d363c75444ada194cc0d47fecd3d80c132 WatchSource:0}: Error finding container 2067c9beafc471fc2df3c09f452916d363c75444ada194cc0d47fecd3d80c132: Status 404 returned error can't find the container with id 2067c9beafc471fc2df3c09f452916d363c75444ada194cc0d47fecd3d80c132 Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.499669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerStarted","Data":"2067c9beafc471fc2df3c09f452916d363c75444ada194cc0d47fecd3d80c132"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.502559 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26xd2" event={"ID":"69e3f17e-3310-4ef3-8de1-637bf905a126","Type":"ContainerStarted","Data":"c792bbffae1fc447430057e8a11d5f3b661d7e5ff055b189341bb1101a2d6529"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.502615 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26xd2" event={"ID":"69e3f17e-3310-4ef3-8de1-637bf905a126","Type":"ContainerStarted","Data":"c6b4d952603be2a02010166aff2fc5955c87268a276ee229c9315ad377278e6b"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.503714 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.530868 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-26xd2" podStartSLOduration=2.530849051 podStartE2EDuration="2.530849051s" podCreationTimestamp="2025-12-01 20:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:32.52631216 +0000 UTC m=+1327.813803819" watchObservedRunningTime="2025-12-01 20:01:32.530849051 +0000 UTC m=+1327.818340720" Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.575999 4960 generic.go:334] "Generic (PLEG): container finished" podID="207ed224-6521-4483-80cb-be6014a13564" containerID="873b94242f3210e3c54d5794879feae58ed646c6e7bb493d35f615bdd1ab6e68" exitCode=0 Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.576100 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" event={"ID":"207ed224-6521-4483-80cb-be6014a13564","Type":"ContainerDied","Data":"873b94242f3210e3c54d5794879feae58ed646c6e7bb493d35f615bdd1ab6e68"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.576140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" event={"ID":"207ed224-6521-4483-80cb-be6014a13564","Type":"ContainerStarted","Data":"993fb36cb5e4da0307e006336ed3e12d3b4bacbe472bb32f914f54711f3c0c3f"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.589193 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.592234 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qmjlr" event={"ID":"4a9d0791-d18b-4346-be7b-007e09b3add7","Type":"ContainerStarted","Data":"35067c17252591b0bbba800079cca55849af10bd5f0ad5506575f70bef3ec0fb"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.601414 4960 generic.go:334] "Generic (PLEG): container finished" podID="40de8a90-fdbc-4a2a-974f-d3e7d266af31" containerID="9f919dca7fdfd3af81a26c11eecbea26de7d2cfe5cad3e2745703c3a96103977" exitCode=0 Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.601500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" event={"ID":"40de8a90-fdbc-4a2a-974f-d3e7d266af31","Type":"ContainerDied","Data":"9f919dca7fdfd3af81a26c11eecbea26de7d2cfe5cad3e2745703c3a96103977"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.641803 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerStarted","Data":"6b59ddb4bb1454102201333953de118de892cf1fbc4816a5d08c80c68f5f322b"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.679309 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6rmd" event={"ID":"e413a5b2-dcf3-4179-9880-b2cec1a2ef56","Type":"ContainerStarted","Data":"87780ef164b901da83309a5993da3e334b0c717350640ef9f5375c7f045a6a38"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.714450 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.719476 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gm7vf" event={"ID":"081b4a0b-4b47-4daa-a407-0e9076b71363","Type":"ContainerStarted","Data":"b67f672d0c21fd16cc2d41ae3278ab79a01c86ded2c7e9dde99c984294809c3e"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.739186 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zdrn9" event={"ID":"80b90b04-9bf0-4bab-a581-4691ef53713e","Type":"ContainerStarted","Data":"faec59cfbf7a7299d1a6802b84ee439c70216a7129d8d046fdc5a0c674b141db"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.755336 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nd25g" event={"ID":"99e4eb83-248f-4900-95fc-ea628cfcf067","Type":"ContainerStarted","Data":"73c644f8ee7986d07b1f3b0f13490204e33643fb0c66585a4b511832f0b7471b"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.763611 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerStarted","Data":"9cef3201d3ad46d40d8fe97c1b98560e83a53c65cf8425300e8a9ffe432661f2"} Dec 01 20:01:32 crc kubenswrapper[4960]: I1201 20:01:32.786796 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-gm7vf" podStartSLOduration=3.786773413 podStartE2EDuration="3.786773413s" podCreationTimestamp="2025-12-01 20:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:32.751259193 +0000 UTC m=+1328.038750872" watchObservedRunningTime="2025-12-01 20:01:32.786773413 +0000 UTC m=+1328.074265082" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.329538 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.409804 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.409911 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.410053 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9hxl\" (UniqueName: \"kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.410089 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.410168 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.410200 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc\") pod \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\" (UID: \"40de8a90-fdbc-4a2a-974f-d3e7d266af31\") " Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.476381 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl" (OuterVolumeSpecName: "kube-api-access-g9hxl") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "kube-api-access-g9hxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.509751 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.513484 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9hxl\" (UniqueName: \"kubernetes.io/projected/40de8a90-fdbc-4a2a-974f-d3e7d266af31-kube-api-access-g9hxl\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.513517 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.573070 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.622544 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.682445 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config" (OuterVolumeSpecName: "config") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.693612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.697690 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "40de8a90-fdbc-4a2a-974f-d3e7d266af31" (UID: "40de8a90-fdbc-4a2a-974f-d3e7d266af31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.724167 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.724216 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.724228 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40de8a90-fdbc-4a2a-974f-d3e7d266af31-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.811870 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" event={"ID":"207ed224-6521-4483-80cb-be6014a13564","Type":"ContainerStarted","Data":"c6a6e83516d0a62542f517060fc227d84c8185a4dd9beb66893bfc0b0418441f"} Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.812838 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.818055 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.823194 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-l7rjv" event={"ID":"40de8a90-fdbc-4a2a-974f-d3e7d266af31","Type":"ContainerDied","Data":"24780b5cb870b2bf602327de889ec5325763f142706015f6336b6333d11afe6f"} Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.823267 4960 scope.go:117] "RemoveContainer" containerID="9f919dca7fdfd3af81a26c11eecbea26de7d2cfe5cad3e2745703c3a96103977" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.849873 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" podStartSLOduration=3.849854599 podStartE2EDuration="3.849854599s" podCreationTimestamp="2025-12-01 20:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:33.833257486 +0000 UTC m=+1329.120749175" watchObservedRunningTime="2025-12-01 20:01:33.849854599 +0000 UTC m=+1329.137346378" Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.953843 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:33 crc kubenswrapper[4960]: I1201 20:01:33.974660 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-l7rjv"] Dec 01 20:01:34 crc kubenswrapper[4960]: I1201 20:01:34.834827 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerStarted","Data":"c4937f5c5345950a716e7fa938235c46e8f45ee0dae8bd5d895a6c573a1365cb"} Dec 01 20:01:34 crc kubenswrapper[4960]: I1201 20:01:34.842965 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerStarted","Data":"5a3c2f66053bd24ad38dd4747749d0ba7e3bed8995380cc1196c3557fc5cfd7b"} Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.343231 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40de8a90-fdbc-4a2a-974f-d3e7d266af31" path="/var/lib/kubelet/pods/40de8a90-fdbc-4a2a-974f-d3e7d266af31/volumes" Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.895362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerStarted","Data":"d86b21191e3ca5aa23351fa8deb27b0be429eb2392f5b73c5b686da85237a203"} Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.895516 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-log" containerID="cri-o://5a3c2f66053bd24ad38dd4747749d0ba7e3bed8995380cc1196c3557fc5cfd7b" gracePeriod=30 Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.895963 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-httpd" containerID="cri-o://d86b21191e3ca5aa23351fa8deb27b0be429eb2392f5b73c5b686da85237a203" gracePeriod=30 Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.908341 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-log" containerID="cri-o://c4937f5c5345950a716e7fa938235c46e8f45ee0dae8bd5d895a6c573a1365cb" gracePeriod=30 Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.908789 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerStarted","Data":"3e27ef7712cf7f2c271216089ce8a3010e958a91ec641d0f2da8eb99b55fa97b"} Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.908846 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-httpd" containerID="cri-o://3e27ef7712cf7f2c271216089ce8a3010e958a91ec641d0f2da8eb99b55fa97b" gracePeriod=30 Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.957864 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.9578429 podStartE2EDuration="6.9578429s" podCreationTimestamp="2025-12-01 20:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:35.946420067 +0000 UTC m=+1331.233911756" watchObservedRunningTime="2025-12-01 20:01:35.9578429 +0000 UTC m=+1331.245334569" Dec 01 20:01:35 crc kubenswrapper[4960]: I1201 20:01:35.980520 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.980504922 podStartE2EDuration="6.980504922s" podCreationTimestamp="2025-12-01 20:01:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:01:35.979350436 +0000 UTC m=+1331.266842105" watchObservedRunningTime="2025-12-01 20:01:35.980504922 +0000 UTC m=+1331.267996591" Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.917824 4960 generic.go:334] "Generic (PLEG): container finished" podID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerID="d86b21191e3ca5aa23351fa8deb27b0be429eb2392f5b73c5b686da85237a203" exitCode=0 Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.917854 4960 generic.go:334] "Generic (PLEG): container finished" podID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerID="5a3c2f66053bd24ad38dd4747749d0ba7e3bed8995380cc1196c3557fc5cfd7b" exitCode=143 Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.917867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerDied","Data":"d86b21191e3ca5aa23351fa8deb27b0be429eb2392f5b73c5b686da85237a203"} Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.917919 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerDied","Data":"5a3c2f66053bd24ad38dd4747749d0ba7e3bed8995380cc1196c3557fc5cfd7b"} Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.920097 4960 generic.go:334] "Generic (PLEG): container finished" podID="04346d5b-21df-4897-a458-6b8cde19298d" containerID="3e27ef7712cf7f2c271216089ce8a3010e958a91ec641d0f2da8eb99b55fa97b" exitCode=0 Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.920131 4960 generic.go:334] "Generic (PLEG): container finished" podID="04346d5b-21df-4897-a458-6b8cde19298d" containerID="c4937f5c5345950a716e7fa938235c46e8f45ee0dae8bd5d895a6c573a1365cb" exitCode=143 Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.920149 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerDied","Data":"3e27ef7712cf7f2c271216089ce8a3010e958a91ec641d0f2da8eb99b55fa97b"} Dec 01 20:01:36 crc kubenswrapper[4960]: I1201 20:01:36.920179 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerDied","Data":"c4937f5c5345950a716e7fa938235c46e8f45ee0dae8bd5d895a6c573a1365cb"} Dec 01 20:01:37 crc kubenswrapper[4960]: I1201 20:01:37.931603 4960 generic.go:334] "Generic (PLEG): container finished" podID="081b4a0b-4b47-4daa-a407-0e9076b71363" containerID="b67f672d0c21fd16cc2d41ae3278ab79a01c86ded2c7e9dde99c984294809c3e" exitCode=0 Dec 01 20:01:37 crc kubenswrapper[4960]: I1201 20:01:37.931670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gm7vf" event={"ID":"081b4a0b-4b47-4daa-a407-0e9076b71363","Type":"ContainerDied","Data":"b67f672d0c21fd16cc2d41ae3278ab79a01c86ded2c7e9dde99c984294809c3e"} Dec 01 20:01:40 crc kubenswrapper[4960]: I1201 20:01:40.608509 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:01:40 crc kubenswrapper[4960]: I1201 20:01:40.696222 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:40 crc kubenswrapper[4960]: I1201 20:01:40.696472 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" containerID="cri-o://b4bfe30915785ce3ecbdb5707d2f4bd7bd92a15c06e4f8a7e4fb4fa71b8992e5" gracePeriod=10 Dec 01 20:01:40 crc kubenswrapper[4960]: I1201 20:01:40.969066 4960 generic.go:334] "Generic (PLEG): container finished" podID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerID="b4bfe30915785ce3ecbdb5707d2f4bd7bd92a15c06e4f8a7e4fb4fa71b8992e5" exitCode=0 Dec 01 20:01:40 crc kubenswrapper[4960]: I1201 20:01:40.969128 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" event={"ID":"960cddd2-52b0-4c62-b45f-fca88b7b6183","Type":"ContainerDied","Data":"b4bfe30915785ce3ecbdb5707d2f4bd7bd92a15c06e4f8a7e4fb4fa71b8992e5"} Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.136002 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.144585 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246313 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246470 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246504 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vhxr\" (UniqueName: \"kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246560 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246591 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246621 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246678 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246707 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246736 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhrbk\" (UniqueName: \"kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246792 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246854 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.246959 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.247021 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"04346d5b-21df-4897-a458-6b8cde19298d\" (UID: \"04346d5b-21df-4897-a458-6b8cde19298d\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.247161 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts\") pod \"081b4a0b-4b47-4daa-a407-0e9076b71363\" (UID: \"081b4a0b-4b47-4daa-a407-0e9076b71363\") " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.247676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs" (OuterVolumeSpecName: "logs") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.247892 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.247907 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04346d5b-21df-4897-a458-6b8cde19298d-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.257303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr" (OuterVolumeSpecName: "kube-api-access-9vhxr") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "kube-api-access-9vhxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.260142 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.265980 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts" (OuterVolumeSpecName: "scripts") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.267424 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk" (OuterVolumeSpecName: "kube-api-access-vhrbk") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "kube-api-access-vhrbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.270237 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts" (OuterVolumeSpecName: "scripts") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.273330 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.283728 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0" (OuterVolumeSpecName: "glance") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.289301 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data" (OuterVolumeSpecName: "config-data") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.297146 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.303313 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "081b4a0b-4b47-4daa-a407-0e9076b71363" (UID: "081b4a0b-4b47-4daa-a407-0e9076b71363"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.331713 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data" (OuterVolumeSpecName: "config-data") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.332173 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "04346d5b-21df-4897-a458-6b8cde19298d" (UID: "04346d5b-21df-4897-a458-6b8cde19298d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.349997 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350035 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350051 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350060 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vhxr\" (UniqueName: \"kubernetes.io/projected/081b4a0b-4b47-4daa-a407-0e9076b71363-kube-api-access-9vhxr\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350070 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350080 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350088 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350096 4960 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/081b4a0b-4b47-4daa-a407-0e9076b71363-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350107 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhrbk\" (UniqueName: \"kubernetes.io/projected/04346d5b-21df-4897-a458-6b8cde19298d-kube-api-access-vhrbk\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350128 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350136 4960 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04346d5b-21df-4897-a458-6b8cde19298d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.350172 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") on node \"crc\" " Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.398821 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.399043 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0") on node "crc" Dec 01 20:01:42 crc kubenswrapper[4960]: I1201 20:01:42.451992 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.026104 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gm7vf" event={"ID":"081b4a0b-4b47-4daa-a407-0e9076b71363","Type":"ContainerDied","Data":"c8e6f4d14686b9e39bddd3314e323b354a13e3c0afacfe894ea8c64a1a2d609e"} Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.026141 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gm7vf" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.026158 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8e6f4d14686b9e39bddd3314e323b354a13e3c0afacfe894ea8c64a1a2d609e" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.029006 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04346d5b-21df-4897-a458-6b8cde19298d","Type":"ContainerDied","Data":"2067c9beafc471fc2df3c09f452916d363c75444ada194cc0d47fecd3d80c132"} Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.029046 4960 scope.go:117] "RemoveContainer" containerID="3e27ef7712cf7f2c271216089ce8a3010e958a91ec641d0f2da8eb99b55fa97b" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.029063 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.070742 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.091622 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108155 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:43 crc kubenswrapper[4960]: E1201 20:01:43.108611 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081b4a0b-4b47-4daa-a407-0e9076b71363" containerName="keystone-bootstrap" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108627 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="081b4a0b-4b47-4daa-a407-0e9076b71363" containerName="keystone-bootstrap" Dec 01 20:01:43 crc kubenswrapper[4960]: E1201 20:01:43.108636 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-httpd" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108645 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-httpd" Dec 01 20:01:43 crc kubenswrapper[4960]: E1201 20:01:43.108668 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40de8a90-fdbc-4a2a-974f-d3e7d266af31" containerName="init" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108673 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="40de8a90-fdbc-4a2a-974f-d3e7d266af31" containerName="init" Dec 01 20:01:43 crc kubenswrapper[4960]: E1201 20:01:43.108686 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-log" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108692 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-log" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108875 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-httpd" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108888 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="40de8a90-fdbc-4a2a-974f-d3e7d266af31" containerName="init" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108901 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="04346d5b-21df-4897-a458-6b8cde19298d" containerName="glance-log" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.108916 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="081b4a0b-4b47-4daa-a407-0e9076b71363" containerName="keystone-bootstrap" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.110061 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.113763 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.113917 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.121615 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.168859 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.168914 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.168992 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.169038 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.169197 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.169295 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hdq\" (UniqueName: \"kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.169377 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.169447 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271130 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271200 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271276 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271320 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271414 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hdq\" (UniqueName: \"kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271469 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.271513 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.272137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.277272 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.277776 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.277809 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/720c6baa92eb97844a8354a6d35963468f1db26ad29581ee9ba5e8c56ff873e5/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.278560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.278563 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.280394 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.281777 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.296886 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hdq\" (UniqueName: \"kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.313216 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-gm7vf"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.321225 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.337959 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04346d5b-21df-4897-a458-6b8cde19298d" path="/var/lib/kubelet/pods/04346d5b-21df-4897-a458-6b8cde19298d/volumes" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.338578 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-gm7vf"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.401201 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bt929"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.402509 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.405174 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mm9q9" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.405271 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.405174 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.405799 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.411395 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bt929"] Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.430388 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.475415 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.475529 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.475734 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.475767 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.476333 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.476403 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q4zm\" (UniqueName: \"kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578363 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578414 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q4zm\" (UniqueName: \"kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578434 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578484 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.578524 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.581975 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.582321 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.582572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.583128 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.583600 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.599596 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q4zm\" (UniqueName: \"kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm\") pod \"keystone-bootstrap-bt929\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.733654 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bt929" Dec 01 20:01:43 crc kubenswrapper[4960]: I1201 20:01:43.847528 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: connect: connection refused" Dec 01 20:01:45 crc kubenswrapper[4960]: I1201 20:01:45.342381 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081b4a0b-4b47-4daa-a407-0e9076b71363" path="/var/lib/kubelet/pods/081b4a0b-4b47-4daa-a407-0e9076b71363/volumes" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.521755 4960 scope.go:117] "RemoveContainer" containerID="c4937f5c5345950a716e7fa938235c46e8f45ee0dae8bd5d895a6c573a1365cb" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.612581 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.714882 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.714974 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715055 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715206 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715240 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715313 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715413 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78v99\" (UniqueName: \"kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.715449 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts\") pod \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\" (UID: \"036dc67e-619f-46e2-aec9-97c2ee5c74a7\") " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.717408 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs" (OuterVolumeSpecName: "logs") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.718580 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.722787 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts" (OuterVolumeSpecName: "scripts") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.729133 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99" (OuterVolumeSpecName: "kube-api-access-78v99") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "kube-api-access-78v99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.730400 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52" (OuterVolumeSpecName: "glance") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.749702 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.769502 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.791369 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data" (OuterVolumeSpecName: "config-data") pod "036dc67e-619f-46e2-aec9-97c2ee5c74a7" (UID: "036dc67e-619f-46e2-aec9-97c2ee5c74a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818240 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818280 4960 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818316 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") on node \"crc\" " Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818327 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818343 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/036dc67e-619f-46e2-aec9-97c2ee5c74a7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818350 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78v99\" (UniqueName: \"kubernetes.io/projected/036dc67e-619f-46e2-aec9-97c2ee5c74a7-kube-api-access-78v99\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818360 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.818367 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/036dc67e-619f-46e2-aec9-97c2ee5c74a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.839353 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.839728 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52") on node "crc" Dec 01 20:01:49 crc kubenswrapper[4960]: I1201 20:01:49.920414 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.104262 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.104258 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"036dc67e-619f-46e2-aec9-97c2ee5c74a7","Type":"ContainerDied","Data":"9cef3201d3ad46d40d8fe97c1b98560e83a53c65cf8425300e8a9ffe432661f2"} Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.141813 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.151414 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.177964 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:50 crc kubenswrapper[4960]: E1201 20:01:50.178638 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-httpd" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.178654 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-httpd" Dec 01 20:01:50 crc kubenswrapper[4960]: E1201 20:01:50.178698 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-log" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.178704 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-log" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.178901 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-httpd" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.178916 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" containerName="glance-log" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.181006 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.184807 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.185014 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.195161 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334218 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bpgx\" (UniqueName: \"kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334260 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334301 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334335 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334355 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334575 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.334619 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.435880 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.435953 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.435996 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436017 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436107 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436192 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436228 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bpgx\" (UniqueName: \"kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436674 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.436705 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.438800 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.438830 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91baf06d123dd91aed63649f36b06b6d82695580db5bb21a6dba903e67f7a922/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.440205 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.440307 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.441906 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.442377 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.457811 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bpgx\" (UniqueName: \"kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.478751 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " pod="openstack/glance-default-external-api-0" Dec 01 20:01:50 crc kubenswrapper[4960]: I1201 20:01:50.502459 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:01:51 crc kubenswrapper[4960]: I1201 20:01:51.339970 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036dc67e-619f-46e2-aec9-97c2ee5c74a7" path="/var/lib/kubelet/pods/036dc67e-619f-46e2-aec9-97c2ee5c74a7/volumes" Dec 01 20:01:52 crc kubenswrapper[4960]: I1201 20:01:52.134623 4960 generic.go:334] "Generic (PLEG): container finished" podID="69e3f17e-3310-4ef3-8de1-637bf905a126" containerID="c792bbffae1fc447430057e8a11d5f3b661d7e5ff055b189341bb1101a2d6529" exitCode=0 Dec 01 20:01:52 crc kubenswrapper[4960]: I1201 20:01:52.134669 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26xd2" event={"ID":"69e3f17e-3310-4ef3-8de1-637bf905a126","Type":"ContainerDied","Data":"c792bbffae1fc447430057e8a11d5f3b661d7e5ff055b189341bb1101a2d6529"} Dec 01 20:01:53 crc kubenswrapper[4960]: I1201 20:01:53.847533 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: i/o timeout" Dec 01 20:01:58 crc kubenswrapper[4960]: E1201 20:01:58.663251 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 01 20:01:58 crc kubenswrapper[4960]: E1201 20:01:58.663751 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65dhd6hc4h699hf7hf8h568h696h67fh598h568h576h68ch5ddh8fh679h654h5d7h58h589h698h668hdch56h659h685h5b4hd5hd5hd6h5b6h5d4q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2272f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(e7f05fc9-a84c-4ae2-97df-439c5fccf1cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.790013 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.848924 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: i/o timeout" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.849029 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919268 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919402 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919450 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919484 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919563 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.919739 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxrjx\" (UniqueName: \"kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx\") pod \"960cddd2-52b0-4c62-b45f-fca88b7b6183\" (UID: \"960cddd2-52b0-4c62-b45f-fca88b7b6183\") " Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.925028 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx" (OuterVolumeSpecName: "kube-api-access-nxrjx") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "kube-api-access-nxrjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.968138 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.973389 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config" (OuterVolumeSpecName: "config") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.985795 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.986857 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:58 crc kubenswrapper[4960]: I1201 20:01:58.992707 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "960cddd2-52b0-4c62-b45f-fca88b7b6183" (UID: "960cddd2-52b0-4c62-b45f-fca88b7b6183"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022513 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022545 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022554 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022563 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022573 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/960cddd2-52b0-4c62-b45f-fca88b7b6183-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.022582 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxrjx\" (UniqueName: \"kubernetes.io/projected/960cddd2-52b0-4c62-b45f-fca88b7b6183-kube-api-access-nxrjx\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.139285 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.201194 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" event={"ID":"960cddd2-52b0-4c62-b45f-fca88b7b6183","Type":"ContainerDied","Data":"b971ca8bb725f3535853ac9ed38d912e6dc5bef00baa75fe105bfea7a6b17537"} Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.201263 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-t42dw" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.204220 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-26xd2" event={"ID":"69e3f17e-3310-4ef3-8de1-637bf905a126","Type":"ContainerDied","Data":"c6b4d952603be2a02010166aff2fc5955c87268a276ee229c9315ad377278e6b"} Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.204258 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6b4d952603be2a02010166aff2fc5955c87268a276ee229c9315ad377278e6b" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.204275 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-26xd2" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.225997 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle\") pod \"69e3f17e-3310-4ef3-8de1-637bf905a126\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.226239 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5knkj\" (UniqueName: \"kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj\") pod \"69e3f17e-3310-4ef3-8de1-637bf905a126\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.232212 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config\") pod \"69e3f17e-3310-4ef3-8de1-637bf905a126\" (UID: \"69e3f17e-3310-4ef3-8de1-637bf905a126\") " Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.252714 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.253761 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj" (OuterVolumeSpecName: "kube-api-access-5knkj") pod "69e3f17e-3310-4ef3-8de1-637bf905a126" (UID: "69e3f17e-3310-4ef3-8de1-637bf905a126"). InnerVolumeSpecName "kube-api-access-5knkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.267447 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-t42dw"] Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.269751 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config" (OuterVolumeSpecName: "config") pod "69e3f17e-3310-4ef3-8de1-637bf905a126" (UID: "69e3f17e-3310-4ef3-8de1-637bf905a126"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.288320 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69e3f17e-3310-4ef3-8de1-637bf905a126" (UID: "69e3f17e-3310-4ef3-8de1-637bf905a126"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.338364 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" path="/var/lib/kubelet/pods/960cddd2-52b0-4c62-b45f-fca88b7b6183/volumes" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.344710 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.344750 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e3f17e-3310-4ef3-8de1-637bf905a126-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:01:59 crc kubenswrapper[4960]: I1201 20:01:59.344761 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5knkj\" (UniqueName: \"kubernetes.io/projected/69e3f17e-3310-4ef3-8de1-637bf905a126-kube-api-access-5knkj\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.355783 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.356654 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="init" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.356673 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="init" Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.356693 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.356700 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.356725 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e3f17e-3310-4ef3-8de1-637bf905a126" containerName="neutron-db-sync" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.356732 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e3f17e-3310-4ef3-8de1-637bf905a126" containerName="neutron-db-sync" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.356955 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e3f17e-3310-4ef3-8de1-637bf905a126" containerName="neutron-db-sync" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.356972 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="960cddd2-52b0-4c62-b45f-fca88b7b6183" containerName="dnsmasq-dns" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.358392 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.368359 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.498713 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.498992 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.499147 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.499263 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.499365 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6prml\" (UniqueName: \"kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.499542 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.510818 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.513726 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.516926 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.517180 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.517298 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jf8bn" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.517423 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.518351 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.518469 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zkqm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-l6rmd_openstack(e413a5b2-dcf3-4179-9880-b2cec1a2ef56): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:02:00 crc kubenswrapper[4960]: E1201 20:02:00.519989 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-l6rmd" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.546457 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602365 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602524 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602615 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602641 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.602667 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6prml\" (UniqueName: \"kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.604064 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.604180 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.604280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.604376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.604632 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.621108 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6prml\" (UniqueName: \"kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml\") pod \"dnsmasq-dns-55f844cf75-5gct5\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.689066 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.704575 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h9dv\" (UniqueName: \"kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.704660 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.704689 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.704748 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.704776 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.805911 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.805972 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.806040 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h9dv\" (UniqueName: \"kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.806096 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.806150 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.811390 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.812040 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.813241 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.824087 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h9dv\" (UniqueName: \"kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.843890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config\") pod \"neutron-dd6798df6-k67nk\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:00 crc kubenswrapper[4960]: I1201 20:02:00.880212 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:01 crc kubenswrapper[4960]: E1201 20:02:01.230392 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-l6rmd" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.681544 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bd769db7-hsw86"] Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.687484 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.696757 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd769db7-hsw86"] Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.707528 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.707870 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860337 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-internal-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860390 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-combined-ca-bundle\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860568 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860615 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-httpd-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860636 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq8pw\" (UniqueName: \"kubernetes.io/projected/aee39569-b67b-419a-ad6a-461bbed0d585-kube-api-access-mq8pw\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860702 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-public-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.860727 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-ovndb-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.962862 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-combined-ca-bundle\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.962954 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.962972 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-httpd-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.962993 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq8pw\" (UniqueName: \"kubernetes.io/projected/aee39569-b67b-419a-ad6a-461bbed0d585-kube-api-access-mq8pw\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.963023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-public-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.963042 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-ovndb-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.963160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-internal-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.968636 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-httpd-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.968694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-combined-ca-bundle\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.969977 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-ovndb-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.976379 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-config\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.977016 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-public-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.982739 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aee39569-b67b-419a-ad6a-461bbed0d585-internal-tls-certs\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:02 crc kubenswrapper[4960]: I1201 20:02:02.986562 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq8pw\" (UniqueName: \"kubernetes.io/projected/aee39569-b67b-419a-ad6a-461bbed0d585-kube-api-access-mq8pw\") pod \"neutron-bd769db7-hsw86\" (UID: \"aee39569-b67b-419a-ad6a-461bbed0d585\") " pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:03 crc kubenswrapper[4960]: I1201 20:02:03.008459 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:03 crc kubenswrapper[4960]: I1201 20:02:03.075149 4960 scope.go:117] "RemoveContainer" containerID="d86b21191e3ca5aa23351fa8deb27b0be429eb2392f5b73c5b686da85237a203" Dec 01 20:02:03 crc kubenswrapper[4960]: I1201 20:02:03.631211 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bt929"] Dec 01 20:02:03 crc kubenswrapper[4960]: I1201 20:02:03.639585 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:02:05 crc kubenswrapper[4960]: W1201 20:02:05.645446 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b4c77e1_7c26_40f7_a01c_1902857fb5eb.slice/crio-61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d WatchSource:0}: Error finding container 61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d: Status 404 returned error can't find the container with id 61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d Dec 01 20:02:05 crc kubenswrapper[4960]: W1201 20:02:05.646091 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37988728_7b89_4f2c_bb2d_979a5339be9e.slice/crio-ee01c02c12f398f06b8b1904767bd31aa41642e39b25b4f84463bf75c3b92923 WatchSource:0}: Error finding container ee01c02c12f398f06b8b1904767bd31aa41642e39b25b4f84463bf75c3b92923: Status 404 returned error can't find the container with id ee01c02c12f398f06b8b1904767bd31aa41642e39b25b4f84463bf75c3b92923 Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.119454 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.245275 4960 scope.go:117] "RemoveContainer" containerID="5a3c2f66053bd24ad38dd4747749d0ba7e3bed8995380cc1196c3557fc5cfd7b" Dec 01 20:02:06 crc kubenswrapper[4960]: W1201 20:02:06.251793 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ed64227_884f_41f4_8534_1bc160d45726.slice/crio-7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e WatchSource:0}: Error finding container 7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e: Status 404 returned error can't find the container with id 7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e Dec 01 20:02:06 crc kubenswrapper[4960]: E1201 20:02:06.271808 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 01 20:02:06 crc kubenswrapper[4960]: E1201 20:02:06.271855 4960 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current" Dec 01 20:02:06 crc kubenswrapper[4960]: E1201 20:02:06.272013 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cloudkitty-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CloudKittyPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:CloudKittyPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:cloudkitty-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:certs,ReadOnly:true,MountPath:/var/lib/openstack/loki-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vgg4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42406,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cloudkitty-db-sync-zdrn9_openstack(80b90b04-9bf0-4bab-a581-4691ef53713e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:02:06 crc kubenswrapper[4960]: E1201 20:02:06.273751 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cloudkitty-db-sync-zdrn9" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.310685 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerStarted","Data":"7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e"} Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.322360 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerStarted","Data":"ee01c02c12f398f06b8b1904767bd31aa41642e39b25b4f84463bf75c3b92923"} Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.324772 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bt929" event={"ID":"2b4c77e1-7c26-40f7-a01c-1902857fb5eb","Type":"ContainerStarted","Data":"61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d"} Dec 01 20:02:06 crc kubenswrapper[4960]: E1201 20:02:06.327935 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cloudkitty-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current\\\"\"" pod="openstack/cloudkitty-db-sync-zdrn9" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.790068 4960 scope.go:117] "RemoveContainer" containerID="b4bfe30915785ce3ecbdb5707d2f4bd7bd92a15c06e4f8a7e4fb4fa71b8992e5" Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.808323 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd769db7-hsw86"] Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.855701 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:06 crc kubenswrapper[4960]: W1201 20:02:06.892879 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod407668ff_7bd5_44c9_ae04_b479c81d45ce.slice/crio-990f44e04d54ca2e3300f55059ffbedf6914cf189ce6e5fcc1e2e3a8d18a8dc9 WatchSource:0}: Error finding container 990f44e04d54ca2e3300f55059ffbedf6914cf189ce6e5fcc1e2e3a8d18a8dc9: Status 404 returned error can't find the container with id 990f44e04d54ca2e3300f55059ffbedf6914cf189ce6e5fcc1e2e3a8d18a8dc9 Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.914622 4960 scope.go:117] "RemoveContainer" containerID="c7af915a692e7cca853ac5179a0bce304581559f2b3855d014b11404e85d792d" Dec 01 20:02:06 crc kubenswrapper[4960]: I1201 20:02:06.935258 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.253759 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.388657 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bt929" event={"ID":"2b4c77e1-7c26-40f7-a01c-1902857fb5eb","Type":"ContainerStarted","Data":"fb5cefceb3d264c13e860d3900207851bdb4a6a3cbaf021c351a134c295c2cbc"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.388951 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerStarted","Data":"c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.388967 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerStarted","Data":"990f44e04d54ca2e3300f55059ffbedf6914cf189ce6e5fcc1e2e3a8d18a8dc9"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.388979 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd769db7-hsw86" event={"ID":"aee39569-b67b-419a-ad6a-461bbed0d585","Type":"ContainerStarted","Data":"ce86bb42048db0345b01557df45aa8892ec2993e74b718cefda52974b51a2cb2"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.388992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd769db7-hsw86" event={"ID":"aee39569-b67b-419a-ad6a-461bbed0d585","Type":"ContainerStarted","Data":"c4e0111d1c4b0b8f6774213d6e1d8a5d0ec88aed2b1dd1dac5855d1ae2e34e56"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.389001 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerStarted","Data":"bf9b914186246729595c9c68869be37b28c1b46159d8474fe3467995257da219"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.389012 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerStarted","Data":"b44771492dc9e8c4bbf00cdbb631f03cbf065ccc4d7ffbfa17a1f0dcc42bce88"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.391998 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nd25g" event={"ID":"99e4eb83-248f-4900-95fc-ea628cfcf067","Type":"ContainerStarted","Data":"428c9e6da42206ad6467059f7724afd36582f2dc42cb5e46af4c1881182cc6d6"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.414995 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerStarted","Data":"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.446948 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bt929" podStartSLOduration=24.446911513 podStartE2EDuration="24.446911513s" podCreationTimestamp="2025-12-01 20:01:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:07.382959097 +0000 UTC m=+1362.670450776" watchObservedRunningTime="2025-12-01 20:02:07.446911513 +0000 UTC m=+1362.734412242" Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.464786 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-nd25g" podStartSLOduration=10.969169662 podStartE2EDuration="38.464767485s" podCreationTimestamp="2025-12-01 20:01:29 +0000 UTC" firstStartedPulling="2025-12-01 20:01:31.637744745 +0000 UTC m=+1326.925236414" lastFinishedPulling="2025-12-01 20:01:59.133342568 +0000 UTC m=+1354.420834237" observedRunningTime="2025-12-01 20:02:07.42463415 +0000 UTC m=+1362.712125829" watchObservedRunningTime="2025-12-01 20:02:07.464767485 +0000 UTC m=+1362.752259154" Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.470732 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qmjlr" event={"ID":"4a9d0791-d18b-4346-be7b-007e09b3add7","Type":"ContainerStarted","Data":"733ed343a652af54cd2e8f4844a72cc95ad706f139453766e5e4e780a5728776"} Dec 01 20:02:07 crc kubenswrapper[4960]: I1201 20:02:07.501656 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qmjlr" podStartSLOduration=9.534051998 podStartE2EDuration="38.501638435s" podCreationTimestamp="2025-12-01 20:01:29 +0000 UTC" firstStartedPulling="2025-12-01 20:01:31.506452581 +0000 UTC m=+1326.793944250" lastFinishedPulling="2025-12-01 20:02:00.474039018 +0000 UTC m=+1355.761530687" observedRunningTime="2025-12-01 20:02:07.4977408 +0000 UTC m=+1362.785232479" watchObservedRunningTime="2025-12-01 20:02:07.501638435 +0000 UTC m=+1362.789130094" Dec 01 20:02:07 crc kubenswrapper[4960]: E1201 20:02:07.648778 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod407668ff_7bd5_44c9_ae04_b479c81d45ce.slice/crio-conmon-c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod407668ff_7bd5_44c9_ae04_b479c81d45ce.slice/crio-c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd.scope\": RecentStats: unable to find data in memory cache]" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.486520 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd769db7-hsw86" event={"ID":"aee39569-b67b-419a-ad6a-461bbed0d585","Type":"ContainerStarted","Data":"ace449d512e21aefa74ffe65c2634cbb5afbd9e6c67f7e1c94e97884c6dcbf73"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.487249 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.491874 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerStarted","Data":"def26066f8c718e9c7360c95a45942c4818705caa17fa1ccc96fee88aec073f7"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.491918 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerStarted","Data":"8fd8e20abe45a1b5aa11c7209951adfdad2e47d20e9cb8f61bb109149a87f0b7"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.494246 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerStarted","Data":"1d8f606812105aa624867384d53a722d0010b121426d44171d1d0b7694081716"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.494280 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerStarted","Data":"f9b86a25e5c116940aa360b7d179569e5c387f9371bca29da93408cef36aa980"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.511582 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerStarted","Data":"a5d6130b85a0bd978c56b6b903e0e903220cba1cc09016b3f744f6a4846fddb8"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.511712 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.524205 4960 generic.go:334] "Generic (PLEG): container finished" podID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerID="c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd" exitCode=0 Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.527438 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerDied","Data":"c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.531348 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerStarted","Data":"3a2d2dccf519aed23563979e8960e3fb65053dc0fba6f27500aaf3afa6b41645"} Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.531482 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.533613 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bd769db7-hsw86" podStartSLOduration=6.533586791 podStartE2EDuration="6.533586791s" podCreationTimestamp="2025-12-01 20:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:08.513738696 +0000 UTC m=+1363.801230375" watchObservedRunningTime="2025-12-01 20:02:08.533586791 +0000 UTC m=+1363.821078470" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.573022 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=25.573002093 podStartE2EDuration="25.573002093s" podCreationTimestamp="2025-12-01 20:01:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:08.544387647 +0000 UTC m=+1363.831879326" watchObservedRunningTime="2025-12-01 20:02:08.573002093 +0000 UTC m=+1363.860493762" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.605075 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dd6798df6-k67nk" podStartSLOduration=8.605049298 podStartE2EDuration="8.605049298s" podCreationTimestamp="2025-12-01 20:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:08.570682569 +0000 UTC m=+1363.858174248" watchObservedRunningTime="2025-12-01 20:02:08.605049298 +0000 UTC m=+1363.892540967" Dec 01 20:02:08 crc kubenswrapper[4960]: I1201 20:02:08.615661 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" podStartSLOduration=8.615623947 podStartE2EDuration="8.615623947s" podCreationTimestamp="2025-12-01 20:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:08.595597176 +0000 UTC m=+1363.883088835" watchObservedRunningTime="2025-12-01 20:02:08.615623947 +0000 UTC m=+1363.903115616" Dec 01 20:02:09 crc kubenswrapper[4960]: I1201 20:02:09.567065 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.567047766 podStartE2EDuration="19.567047766s" podCreationTimestamp="2025-12-01 20:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:09.563983498 +0000 UTC m=+1364.851475167" watchObservedRunningTime="2025-12-01 20:02:09.567047766 +0000 UTC m=+1364.854539435" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.502638 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.503002 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.538622 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.562546 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.568027 4960 generic.go:334] "Generic (PLEG): container finished" podID="4a9d0791-d18b-4346-be7b-007e09b3add7" containerID="733ed343a652af54cd2e8f4844a72cc95ad706f139453766e5e4e780a5728776" exitCode=0 Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.568362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qmjlr" event={"ID":"4a9d0791-d18b-4346-be7b-007e09b3add7","Type":"ContainerDied","Data":"733ed343a652af54cd2e8f4844a72cc95ad706f139453766e5e4e780a5728776"} Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.571468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 20:02:10 crc kubenswrapper[4960]: I1201 20:02:10.571561 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 20:02:11 crc kubenswrapper[4960]: I1201 20:02:11.578724 4960 generic.go:334] "Generic (PLEG): container finished" podID="2b4c77e1-7c26-40f7-a01c-1902857fb5eb" containerID="fb5cefceb3d264c13e860d3900207851bdb4a6a3cbaf021c351a134c295c2cbc" exitCode=0 Dec 01 20:02:11 crc kubenswrapper[4960]: I1201 20:02:11.578812 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bt929" event={"ID":"2b4c77e1-7c26-40f7-a01c-1902857fb5eb","Type":"ContainerDied","Data":"fb5cefceb3d264c13e860d3900207851bdb4a6a3cbaf021c351a134c295c2cbc"} Dec 01 20:02:11 crc kubenswrapper[4960]: I1201 20:02:11.580907 4960 generic.go:334] "Generic (PLEG): container finished" podID="99e4eb83-248f-4900-95fc-ea628cfcf067" containerID="428c9e6da42206ad6467059f7724afd36582f2dc42cb5e46af4c1881182cc6d6" exitCode=0 Dec 01 20:02:11 crc kubenswrapper[4960]: I1201 20:02:11.580982 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nd25g" event={"ID":"99e4eb83-248f-4900-95fc-ea628cfcf067","Type":"ContainerDied","Data":"428c9e6da42206ad6467059f7724afd36582f2dc42cb5e46af4c1881182cc6d6"} Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.434790 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.435351 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.435364 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.435376 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.465105 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:13 crc kubenswrapper[4960]: I1201 20:02:13.477887 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:14 crc kubenswrapper[4960]: I1201 20:02:14.690450 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 20:02:14 crc kubenswrapper[4960]: I1201 20:02:14.702802 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 20:02:15 crc kubenswrapper[4960]: I1201 20:02:15.691936 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:15 crc kubenswrapper[4960]: I1201 20:02:15.749904 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:02:15 crc kubenswrapper[4960]: I1201 20:02:15.750310 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="dnsmasq-dns" containerID="cri-o://c6a6e83516d0a62542f517060fc227d84c8185a4dd9beb66893bfc0b0418441f" gracePeriod=10 Dec 01 20:02:15 crc kubenswrapper[4960]: I1201 20:02:15.918044 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:15 crc kubenswrapper[4960]: I1201 20:02:15.918237 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 20:02:16 crc kubenswrapper[4960]: I1201 20:02:16.127796 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 20:02:17 crc kubenswrapper[4960]: I1201 20:02:17.641256 4960 generic.go:334] "Generic (PLEG): container finished" podID="207ed224-6521-4483-80cb-be6014a13564" containerID="c6a6e83516d0a62542f517060fc227d84c8185a4dd9beb66893bfc0b0418441f" exitCode=0 Dec 01 20:02:17 crc kubenswrapper[4960]: I1201 20:02:17.641606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" event={"ID":"207ed224-6521-4483-80cb-be6014a13564","Type":"ContainerDied","Data":"c6a6e83516d0a62542f517060fc227d84c8185a4dd9beb66893bfc0b0418441f"} Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.606471 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.160:5353: connect: connection refused" Dec 01 20:02:20 crc kubenswrapper[4960]: E1201 20:02:20.743899 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Dec 01 20:02:20 crc kubenswrapper[4960]: E1201 20:02:20.744042 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2272f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(e7f05fc9-a84c-4ae2-97df-439c5fccf1cf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.839953 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qmjlr" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.876265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs\") pod \"4a9d0791-d18b-4346-be7b-007e09b3add7\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.876367 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdpjl\" (UniqueName: \"kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl\") pod \"4a9d0791-d18b-4346-be7b-007e09b3add7\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.876419 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle\") pod \"4a9d0791-d18b-4346-be7b-007e09b3add7\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.876512 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts\") pod \"4a9d0791-d18b-4346-be7b-007e09b3add7\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.876626 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data\") pod \"4a9d0791-d18b-4346-be7b-007e09b3add7\" (UID: \"4a9d0791-d18b-4346-be7b-007e09b3add7\") " Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.880088 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs" (OuterVolumeSpecName: "logs") pod "4a9d0791-d18b-4346-be7b-007e09b3add7" (UID: "4a9d0791-d18b-4346-be7b-007e09b3add7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.889002 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts" (OuterVolumeSpecName: "scripts") pod "4a9d0791-d18b-4346-be7b-007e09b3add7" (UID: "4a9d0791-d18b-4346-be7b-007e09b3add7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.921854 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl" (OuterVolumeSpecName: "kube-api-access-sdpjl") pod "4a9d0791-d18b-4346-be7b-007e09b3add7" (UID: "4a9d0791-d18b-4346-be7b-007e09b3add7"). InnerVolumeSpecName "kube-api-access-sdpjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.978294 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data" (OuterVolumeSpecName: "config-data") pod "4a9d0791-d18b-4346-be7b-007e09b3add7" (UID: "4a9d0791-d18b-4346-be7b-007e09b3add7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.979772 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a9d0791-d18b-4346-be7b-007e09b3add7-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.979793 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdpjl\" (UniqueName: \"kubernetes.io/projected/4a9d0791-d18b-4346-be7b-007e09b3add7-kube-api-access-sdpjl\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.979806 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.979817 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:20 crc kubenswrapper[4960]: I1201 20:02:20.988686 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a9d0791-d18b-4346-be7b-007e09b3add7" (UID: "4a9d0791-d18b-4346-be7b-007e09b3add7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.003389 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bt929" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.018911 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nd25g" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085060 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085144 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q4zm\" (UniqueName: \"kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085196 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data\") pod \"99e4eb83-248f-4900-95fc-ea628cfcf067\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085251 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085286 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085397 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085457 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts\") pod \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\" (UID: \"2b4c77e1-7c26-40f7-a01c-1902857fb5eb\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085558 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfrpt\" (UniqueName: \"kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt\") pod \"99e4eb83-248f-4900-95fc-ea628cfcf067\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.085597 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle\") pod \"99e4eb83-248f-4900-95fc-ea628cfcf067\" (UID: \"99e4eb83-248f-4900-95fc-ea628cfcf067\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.086191 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a9d0791-d18b-4346-be7b-007e09b3add7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.107281 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.118893 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.118974 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "99e4eb83-248f-4900-95fc-ea628cfcf067" (UID: "99e4eb83-248f-4900-95fc-ea628cfcf067"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.119099 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm" (OuterVolumeSpecName: "kube-api-access-2q4zm") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "kube-api-access-2q4zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.127394 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt" (OuterVolumeSpecName: "kube-api-access-rfrpt") pod "99e4eb83-248f-4900-95fc-ea628cfcf067" (UID: "99e4eb83-248f-4900-95fc-ea628cfcf067"). InnerVolumeSpecName "kube-api-access-rfrpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.134234 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts" (OuterVolumeSpecName: "scripts") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.171957 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99e4eb83-248f-4900-95fc-ea628cfcf067" (UID: "99e4eb83-248f-4900-95fc-ea628cfcf067"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187592 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187623 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfrpt\" (UniqueName: \"kubernetes.io/projected/99e4eb83-248f-4900-95fc-ea628cfcf067-kube-api-access-rfrpt\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187637 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187645 4960 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187653 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q4zm\" (UniqueName: \"kubernetes.io/projected/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-kube-api-access-2q4zm\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187662 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/99e4eb83-248f-4900-95fc-ea628cfcf067-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.187671 4960 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.190264 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data" (OuterVolumeSpecName: "config-data") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.204732 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b4c77e1-7c26-40f7-a01c-1902857fb5eb" (UID: "2b4c77e1-7c26-40f7-a01c-1902857fb5eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.206041 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289352 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hshz\" (UniqueName: \"kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289417 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289461 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289536 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289635 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.289690 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0\") pod \"207ed224-6521-4483-80cb-be6014a13564\" (UID: \"207ed224-6521-4483-80cb-be6014a13564\") " Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.290149 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.290167 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b4c77e1-7c26-40f7-a01c-1902857fb5eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.297666 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz" (OuterVolumeSpecName: "kube-api-access-8hshz") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "kube-api-access-8hshz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.347327 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.381997 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config" (OuterVolumeSpecName: "config") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.388649 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.392608 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.392855 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hshz\" (UniqueName: \"kubernetes.io/projected/207ed224-6521-4483-80cb-be6014a13564-kube-api-access-8hshz\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.392932 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.393008 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.400106 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.420809 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "207ed224-6521-4483-80cb-be6014a13564" (UID: "207ed224-6521-4483-80cb-be6014a13564"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.494428 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.494468 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/207ed224-6521-4483-80cb-be6014a13564-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.698283 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" event={"ID":"207ed224-6521-4483-80cb-be6014a13564","Type":"ContainerDied","Data":"993fb36cb5e4da0307e006336ed3e12d3b4bacbe472bb32f914f54711f3c0c3f"} Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.698316 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6p8jz" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.698351 4960 scope.go:117] "RemoveContainer" containerID="c6a6e83516d0a62542f517060fc227d84c8185a4dd9beb66893bfc0b0418441f" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.703827 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qmjlr" event={"ID":"4a9d0791-d18b-4346-be7b-007e09b3add7","Type":"ContainerDied","Data":"35067c17252591b0bbba800079cca55849af10bd5f0ad5506575f70bef3ec0fb"} Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.704089 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35067c17252591b0bbba800079cca55849af10bd5f0ad5506575f70bef3ec0fb" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.704513 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qmjlr" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.718444 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zdrn9" event={"ID":"80b90b04-9bf0-4bab-a581-4691ef53713e","Type":"ContainerStarted","Data":"0ec9baf7f65649b3b8c9c3d9d93cc816bf287dba908ac6b621e15294f8d2eaad"} Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.729661 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nd25g" event={"ID":"99e4eb83-248f-4900-95fc-ea628cfcf067","Type":"ContainerDied","Data":"73c644f8ee7986d07b1f3b0f13490204e33643fb0c66585a4b511832f0b7471b"} Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.729707 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73c644f8ee7986d07b1f3b0f13490204e33643fb0c66585a4b511832f0b7471b" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.729781 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nd25g" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.741970 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bt929" event={"ID":"2b4c77e1-7c26-40f7-a01c-1902857fb5eb","Type":"ContainerDied","Data":"61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d"} Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.742019 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61e9f735d93ccab6e4538d1bd0a838256b1edffe77174357f5e176c3f70d090d" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.742133 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bt929" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.745333 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-zdrn9" podStartSLOduration=2.903633332 podStartE2EDuration="51.745312647s" podCreationTimestamp="2025-12-01 20:01:30 +0000 UTC" firstStartedPulling="2025-12-01 20:01:31.977787631 +0000 UTC m=+1327.265279300" lastFinishedPulling="2025-12-01 20:02:20.819466946 +0000 UTC m=+1376.106958615" observedRunningTime="2025-12-01 20:02:21.739382867 +0000 UTC m=+1377.026874536" watchObservedRunningTime="2025-12-01 20:02:21.745312647 +0000 UTC m=+1377.032804336" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.810231 4960 scope.go:117] "RemoveContainer" containerID="873b94242f3210e3c54d5794879feae58ed646c6e7bb493d35f615bdd1ab6e68" Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.813590 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:02:21 crc kubenswrapper[4960]: I1201 20:02:21.822290 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6p8jz"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.002377 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-74c7bf679b-vg6f9"] Dec 01 20:02:22 crc kubenswrapper[4960]: E1201 20:02:22.002850 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="init" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.002876 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="init" Dec 01 20:02:22 crc kubenswrapper[4960]: E1201 20:02:22.002919 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4c77e1-7c26-40f7-a01c-1902857fb5eb" containerName="keystone-bootstrap" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.002930 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4c77e1-7c26-40f7-a01c-1902857fb5eb" containerName="keystone-bootstrap" Dec 01 20:02:22 crc kubenswrapper[4960]: E1201 20:02:22.002945 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e4eb83-248f-4900-95fc-ea628cfcf067" containerName="barbican-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.002954 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e4eb83-248f-4900-95fc-ea628cfcf067" containerName="barbican-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: E1201 20:02:22.002983 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a9d0791-d18b-4346-be7b-007e09b3add7" containerName="placement-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.002991 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a9d0791-d18b-4346-be7b-007e09b3add7" containerName="placement-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: E1201 20:02:22.003008 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="dnsmasq-dns" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.003017 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="dnsmasq-dns" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.003270 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e4eb83-248f-4900-95fc-ea628cfcf067" containerName="barbican-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.003313 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a9d0791-d18b-4346-be7b-007e09b3add7" containerName="placement-db-sync" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.003327 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="207ed224-6521-4483-80cb-be6014a13564" containerName="dnsmasq-dns" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.003339 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4c77e1-7c26-40f7-a01c-1902857fb5eb" containerName="keystone-bootstrap" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.004744 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.006826 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.006930 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-t4stk" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.007884 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.007916 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.009518 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.020241 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74c7bf679b-vg6f9"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.173572 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5f8c78b6fc-fdfln"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.175034 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.178813 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.178979 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.179033 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.179176 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.179478 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mm9q9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.179656 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214619 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-scripts\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214721 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f8c78b6fc-fdfln"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214805 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-scripts\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214873 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9zt9\" (UniqueName: \"kubernetes.io/projected/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-kube-api-access-z9zt9\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214910 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-internal-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.214992 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs47x\" (UniqueName: \"kubernetes.io/projected/987ef811-8e23-491e-98ff-64fcacab0f37-kube-api-access-gs47x\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-internal-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215056 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-public-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215080 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-credential-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215103 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-logs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215222 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-config-data\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-public-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215540 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-config-data\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215570 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-combined-ca-bundle\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215629 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-combined-ca-bundle\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.215685 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-fernet-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.307001 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f647bc89f-mwqdp"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.308821 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.312158 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-245x8" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.312327 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.312446 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.326943 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f647bc89f-mwqdp"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335097 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-public-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335248 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-config-data\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335272 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-combined-ca-bundle\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335294 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qh7\" (UniqueName: \"kubernetes.io/projected/c6e92e0f-eb53-4d9a-af6a-4294f774a870-kube-api-access-t9qh7\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335329 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-combined-ca-bundle\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335351 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-combined-ca-bundle\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-fernet-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335394 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335424 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-scripts\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335510 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-scripts\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335535 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9zt9\" (UniqueName: \"kubernetes.io/projected/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-kube-api-access-z9zt9\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335553 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-internal-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335587 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs47x\" (UniqueName: \"kubernetes.io/projected/987ef811-8e23-491e-98ff-64fcacab0f37-kube-api-access-gs47x\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335618 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-internal-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335657 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-public-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335676 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-credential-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-logs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335729 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-config-data\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335759 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6e92e0f-eb53-4d9a-af6a-4294f774a870-logs\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.335821 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data-custom\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.347867 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-public-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.353807 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-scripts\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.354603 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-logs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.355432 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-scripts\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.365576 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-internal-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.366157 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-combined-ca-bundle\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.368992 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-config-data\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.372814 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-public-tls-certs\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.374762 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-fernet-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.375324 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-config-data\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.375931 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-internal-tls-certs\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.380750 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9zt9\" (UniqueName: \"kubernetes.io/projected/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-kube-api-access-z9zt9\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.383540 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/987ef811-8e23-491e-98ff-64fcacab0f37-credential-keys\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.383610 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs47x\" (UniqueName: \"kubernetes.io/projected/987ef811-8e23-491e-98ff-64fcacab0f37-kube-api-access-gs47x\") pod \"keystone-5f8c78b6fc-fdfln\" (UID: \"987ef811-8e23-491e-98ff-64fcacab0f37\") " pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.398932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f75be90e-8df9-4f2a-bbdb-6202c7b56cf9-combined-ca-bundle\") pod \"placement-74c7bf679b-vg6f9\" (UID: \"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9\") " pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.420729 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6f67669b46-khzh6"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.422931 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.429916 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.437477 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6e92e0f-eb53-4d9a-af6a-4294f774a870-logs\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.437782 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data-custom\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.437903 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qh7\" (UniqueName: \"kubernetes.io/projected/c6e92e0f-eb53-4d9a-af6a-4294f774a870-kube-api-access-t9qh7\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.437984 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-combined-ca-bundle\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.438078 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.448872 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f67669b46-khzh6"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.453349 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6e92e0f-eb53-4d9a-af6a-4294f774a870-logs\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.454395 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.458293 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-combined-ca-bundle\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.468756 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6e92e0f-eb53-4d9a-af6a-4294f774a870-config-data-custom\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.474898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qh7\" (UniqueName: \"kubernetes.io/projected/c6e92e0f-eb53-4d9a-af6a-4294f774a870-kube-api-access-t9qh7\") pod \"barbican-worker-f647bc89f-mwqdp\" (UID: \"c6e92e0f-eb53-4d9a-af6a-4294f774a870\") " pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.486170 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.487840 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.501608 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.507521 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f647bc89f-mwqdp" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.539399 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.540580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-combined-ca-bundle\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.540711 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prgxb\" (UniqueName: \"kubernetes.io/projected/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-kube-api-access-prgxb\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.540734 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-logs\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.540827 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data-custom\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.540855 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.586203 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.588335 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.590184 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.595653 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.621923 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.642934 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.642985 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-combined-ca-bundle\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.643016 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.643096 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644399 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prgxb\" (UniqueName: \"kubernetes.io/projected/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-kube-api-access-prgxb\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644438 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-logs\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644538 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdrw4\" (UniqueName: \"kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644605 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644627 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data-custom\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644643 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.644666 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.645964 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-logs\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.653558 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data-custom\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.653736 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-combined-ca-bundle\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.654623 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-config-data\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.694702 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prgxb\" (UniqueName: \"kubernetes.io/projected/4771a4c0-c72c-4819-953e-ed6ecd9d1ef6-kube-api-access-prgxb\") pod \"barbican-keystone-listener-6f67669b46-khzh6\" (UID: \"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6\") " pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748263 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8l65\" (UniqueName: \"kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748316 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748395 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748432 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748446 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748483 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748519 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdrw4\" (UniqueName: \"kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748560 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748579 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748600 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.748623 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.749498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.749724 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.750284 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.750525 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.750846 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.782914 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdrw4\" (UniqueName: \"kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4\") pod \"dnsmasq-dns-85ff748b95-5xx8l\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.813408 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6rmd" event={"ID":"e413a5b2-dcf3-4179-9880-b2cec1a2ef56","Type":"ContainerStarted","Data":"5fc63fba6074f60bae235331fdc246083cbde7e8ad97f7c9af74f179c6bf395b"} Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.857239 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8l65\" (UniqueName: \"kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.857309 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.857345 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.857360 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.857398 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.858430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.858446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.863633 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.879962 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-l6rmd" podStartSLOduration=3.9648840979999997 podStartE2EDuration="52.879940509s" podCreationTimestamp="2025-12-01 20:01:30 +0000 UTC" firstStartedPulling="2025-12-01 20:01:31.899499068 +0000 UTC m=+1327.186990737" lastFinishedPulling="2025-12-01 20:02:20.814555479 +0000 UTC m=+1376.102047148" observedRunningTime="2025-12-01 20:02:22.855801417 +0000 UTC m=+1378.143293086" watchObservedRunningTime="2025-12-01 20:02:22.879940509 +0000 UTC m=+1378.167432178" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.882095 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.888263 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.904374 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:22 crc kubenswrapper[4960]: I1201 20:02:22.921422 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8l65\" (UniqueName: \"kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65\") pod \"barbican-api-6877cc4654-npk8m\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.215564 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.366837 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207ed224-6521-4483-80cb-be6014a13564" path="/var/lib/kubelet/pods/207ed224-6521-4483-80cb-be6014a13564/volumes" Dec 01 20:02:23 crc kubenswrapper[4960]: W1201 20:02:23.571243 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6e92e0f_eb53_4d9a_af6a_4294f774a870.slice/crio-34908f22588c3007bf139e9b5bf1664bc1726e3e9a34a8776bc5d56e2377cf9d WatchSource:0}: Error finding container 34908f22588c3007bf139e9b5bf1664bc1726e3e9a34a8776bc5d56e2377cf9d: Status 404 returned error can't find the container with id 34908f22588c3007bf139e9b5bf1664bc1726e3e9a34a8776bc5d56e2377cf9d Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.574470 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f647bc89f-mwqdp"] Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.697514 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5f8c78b6fc-fdfln"] Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.709163 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-74c7bf679b-vg6f9"] Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.864299 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f67669b46-khzh6"] Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.895508 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.944517 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f647bc89f-mwqdp" event={"ID":"c6e92e0f-eb53-4d9a-af6a-4294f774a870","Type":"ContainerStarted","Data":"34908f22588c3007bf139e9b5bf1664bc1726e3e9a34a8776bc5d56e2377cf9d"} Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.953991 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" event={"ID":"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6","Type":"ContainerStarted","Data":"d45d519d85f6e787df8bfbb87e5c802daac7fee1f563a0e075c2e321abd0b3bb"} Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.957247 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c7bf679b-vg6f9" event={"ID":"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9","Type":"ContainerStarted","Data":"686171a828bb0a26d9f09fb0fa52f4fa800b23f9c6dfb43584d20ebccb5b3a5a"} Dec 01 20:02:23 crc kubenswrapper[4960]: I1201 20:02:23.971354 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f8c78b6fc-fdfln" event={"ID":"987ef811-8e23-491e-98ff-64fcacab0f37","Type":"ContainerStarted","Data":"b8efafbc149562a3dcfc936f4f7b91fb07cc46530e504e8672db506cd381a7c9"} Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.183480 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.988937 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c7bf679b-vg6f9" event={"ID":"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9","Type":"ContainerStarted","Data":"5fc90f5a2dd1d2e24663a040ef39eb66135dccc5a2716f6927940f659fc14763"} Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.989502 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-74c7bf679b-vg6f9" event={"ID":"f75be90e-8df9-4f2a-bbdb-6202c7b56cf9","Type":"ContainerStarted","Data":"f88c6b6b7e7c20f252eafbcc840d5c6f5ad4898e863ab53d866bc67ae9ac62c6"} Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.989524 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.989563 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.994016 4960 generic.go:334] "Generic (PLEG): container finished" podID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerID="d8e61c199e3033b465cc14b5e1515159c3cf03214cebfe1ce0d8f1ca93aa8f28" exitCode=0 Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.994670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" event={"ID":"453fc2ad-627e-4256-88f8-6c27cd68211e","Type":"ContainerDied","Data":"d8e61c199e3033b465cc14b5e1515159c3cf03214cebfe1ce0d8f1ca93aa8f28"} Dec 01 20:02:24 crc kubenswrapper[4960]: I1201 20:02:24.994719 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" event={"ID":"453fc2ad-627e-4256-88f8-6c27cd68211e","Type":"ContainerStarted","Data":"3c0b02161d795043b2d643ef24f550ce1d5cf2a886c99a7b342129cee852429c"} Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.000025 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerStarted","Data":"2def03205cf6fc6af7659316ce5aeee330d7d869f904b59a378ee0fa76edec05"} Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.000756 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.000772 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerStarted","Data":"1120540dd51135b7e824078ba3e461c05d58dc1f0991bb8342834740740e88c5"} Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.000786 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.000798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerStarted","Data":"04dba8922d50276f88c48b40f248109476dcd396863bfa309c4041a211acf1dd"} Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.004091 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5f8c78b6fc-fdfln" event={"ID":"987ef811-8e23-491e-98ff-64fcacab0f37","Type":"ContainerStarted","Data":"847a35b81b289e69f55af99d2fdb44d8ff8d37eea931abb3f6858b1420f13d40"} Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.004358 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.035546 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-74c7bf679b-vg6f9" podStartSLOduration=4.035526707 podStartE2EDuration="4.035526707s" podCreationTimestamp="2025-12-01 20:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:25.012608713 +0000 UTC m=+1380.300100372" watchObservedRunningTime="2025-12-01 20:02:25.035526707 +0000 UTC m=+1380.323018376" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.042091 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5f8c78b6fc-fdfln" podStartSLOduration=3.042076136 podStartE2EDuration="3.042076136s" podCreationTimestamp="2025-12-01 20:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:25.040057371 +0000 UTC m=+1380.327549070" watchObservedRunningTime="2025-12-01 20:02:25.042076136 +0000 UTC m=+1380.329567805" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.077859 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6877cc4654-npk8m" podStartSLOduration=3.07784273 podStartE2EDuration="3.07784273s" podCreationTimestamp="2025-12-01 20:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:25.061671993 +0000 UTC m=+1380.349163662" watchObservedRunningTime="2025-12-01 20:02:25.07784273 +0000 UTC m=+1380.365334399" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.795324 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-d89589d9d-9dglh"] Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.797353 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.799073 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.799236 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.816208 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d89589d9d-9dglh"] Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.842737 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk2ln\" (UniqueName: \"kubernetes.io/projected/9c816e6d-aa4e-46cc-8494-39d3f334a028-kube-api-access-pk2ln\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843100 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-internal-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843288 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c816e6d-aa4e-46cc-8494-39d3f334a028-logs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843477 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-public-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843737 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-combined-ca-bundle\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.843899 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data-custom\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.945817 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data-custom\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946210 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk2ln\" (UniqueName: \"kubernetes.io/projected/9c816e6d-aa4e-46cc-8494-39d3f334a028-kube-api-access-pk2ln\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946256 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-internal-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c816e6d-aa4e-46cc-8494-39d3f334a028-logs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-public-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946844 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c816e6d-aa4e-46cc-8494-39d3f334a028-logs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.946892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-combined-ca-bundle\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.952725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-public-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.955234 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-internal-tls-certs\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.956893 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-combined-ca-bundle\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.957588 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data-custom\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.958257 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c816e6d-aa4e-46cc-8494-39d3f334a028-config-data\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:25 crc kubenswrapper[4960]: I1201 20:02:25.967099 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk2ln\" (UniqueName: \"kubernetes.io/projected/9c816e6d-aa4e-46cc-8494-39d3f334a028-kube-api-access-pk2ln\") pod \"barbican-api-d89589d9d-9dglh\" (UID: \"9c816e6d-aa4e-46cc-8494-39d3f334a028\") " pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:26 crc kubenswrapper[4960]: I1201 20:02:26.016100 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" event={"ID":"453fc2ad-627e-4256-88f8-6c27cd68211e","Type":"ContainerStarted","Data":"2dc6c104449fbff3d2ac09a2e4fc67f5471e625b02d903621c8545533d465795"} Dec 01 20:02:26 crc kubenswrapper[4960]: I1201 20:02:26.017056 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:26 crc kubenswrapper[4960]: I1201 20:02:26.043035 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" podStartSLOduration=4.04301963 podStartE2EDuration="4.04301963s" podCreationTimestamp="2025-12-01 20:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:26.041761289 +0000 UTC m=+1381.329252968" watchObservedRunningTime="2025-12-01 20:02:26.04301963 +0000 UTC m=+1381.330511299" Dec 01 20:02:26 crc kubenswrapper[4960]: I1201 20:02:26.130763 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:27 crc kubenswrapper[4960]: I1201 20:02:27.028349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f647bc89f-mwqdp" event={"ID":"c6e92e0f-eb53-4d9a-af6a-4294f774a870","Type":"ContainerStarted","Data":"f1671053f8b9d876f1d82f9a88be19fefd3daf11ea85f4aad8e2a74a0eca5846"} Dec 01 20:02:27 crc kubenswrapper[4960]: I1201 20:02:27.032457 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" event={"ID":"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6","Type":"ContainerStarted","Data":"b8da48095d2d7cedd1dfb29b74f5f3cfd7a1c80ed0b6b7232cdcd34d8e609c68"} Dec 01 20:02:27 crc kubenswrapper[4960]: I1201 20:02:27.060297 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d89589d9d-9dglh"] Dec 01 20:02:27 crc kubenswrapper[4960]: W1201 20:02:27.076390 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c816e6d_aa4e_46cc_8494_39d3f334a028.slice/crio-c8903518f3f3c74c9e3917a81155d2dae30cc608cc734891b4f89477e1f61782 WatchSource:0}: Error finding container c8903518f3f3c74c9e3917a81155d2dae30cc608cc734891b4f89477e1f61782: Status 404 returned error can't find the container with id c8903518f3f3c74c9e3917a81155d2dae30cc608cc734891b4f89477e1f61782 Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.107887 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f647bc89f-mwqdp" event={"ID":"c6e92e0f-eb53-4d9a-af6a-4294f774a870","Type":"ContainerStarted","Data":"cf77a956e55162e24fbc70393b6d922d5d09a9070b6198e4ceb066b176b1a213"} Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.131708 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" event={"ID":"4771a4c0-c72c-4819-953e-ed6ecd9d1ef6","Type":"ContainerStarted","Data":"272d47707c170a7733d01bf52fd05f9b20b50637553e2df62b9ce8dd5c45a930"} Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.168077 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d89589d9d-9dglh" event={"ID":"9c816e6d-aa4e-46cc-8494-39d3f334a028","Type":"ContainerStarted","Data":"2ae546cc9929a2f93aca3b30202fa9fcd4e2367acd0aa35724adf892b2a5e5b4"} Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.168139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d89589d9d-9dglh" event={"ID":"9c816e6d-aa4e-46cc-8494-39d3f334a028","Type":"ContainerStarted","Data":"c8903518f3f3c74c9e3917a81155d2dae30cc608cc734891b4f89477e1f61782"} Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.168572 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f647bc89f-mwqdp" podStartSLOduration=3.183125569 podStartE2EDuration="6.168557324s" podCreationTimestamp="2025-12-01 20:02:22 +0000 UTC" firstStartedPulling="2025-12-01 20:02:23.578152495 +0000 UTC m=+1378.865644164" lastFinishedPulling="2025-12-01 20:02:26.56358425 +0000 UTC m=+1381.851075919" observedRunningTime="2025-12-01 20:02:28.153152722 +0000 UTC m=+1383.440644391" watchObservedRunningTime="2025-12-01 20:02:28.168557324 +0000 UTC m=+1383.456048993" Dec 01 20:02:28 crc kubenswrapper[4960]: I1201 20:02:28.206285 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6f67669b46-khzh6" podStartSLOduration=3.541650804 podStartE2EDuration="6.206268052s" podCreationTimestamp="2025-12-01 20:02:22 +0000 UTC" firstStartedPulling="2025-12-01 20:02:23.890917824 +0000 UTC m=+1379.178409493" lastFinishedPulling="2025-12-01 20:02:26.555535072 +0000 UTC m=+1381.843026741" observedRunningTime="2025-12-01 20:02:28.187860153 +0000 UTC m=+1383.475351822" watchObservedRunningTime="2025-12-01 20:02:28.206268052 +0000 UTC m=+1383.493759721" Dec 01 20:02:30 crc kubenswrapper[4960]: I1201 20:02:30.893041 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:32 crc kubenswrapper[4960]: I1201 20:02:32.210273 4960 generic.go:334] "Generic (PLEG): container finished" podID="80b90b04-9bf0-4bab-a581-4691ef53713e" containerID="0ec9baf7f65649b3b8c9c3d9d93cc816bf287dba908ac6b621e15294f8d2eaad" exitCode=0 Dec 01 20:02:32 crc kubenswrapper[4960]: I1201 20:02:32.210450 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zdrn9" event={"ID":"80b90b04-9bf0-4bab-a581-4691ef53713e","Type":"ContainerDied","Data":"0ec9baf7f65649b3b8c9c3d9d93cc816bf287dba908ac6b621e15294f8d2eaad"} Dec 01 20:02:32 crc kubenswrapper[4960]: I1201 20:02:32.865324 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:32 crc kubenswrapper[4960]: I1201 20:02:32.955986 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:32 crc kubenswrapper[4960]: I1201 20:02:32.956870 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="dnsmasq-dns" containerID="cri-o://3a2d2dccf519aed23563979e8960e3fb65053dc0fba6f27500aaf3afa6b41645" gracePeriod=10 Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.040108 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bd769db7-hsw86" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.140077 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.140370 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-dd6798df6-k67nk" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-api" containerID="cri-o://bf9b914186246729595c9c68869be37b28c1b46159d8474fe3467995257da219" gracePeriod=30 Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.140913 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-dd6798df6-k67nk" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-httpd" containerID="cri-o://a5d6130b85a0bd978c56b6b903e0e903220cba1cc09016b3f744f6a4846fddb8" gracePeriod=30 Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.238599 4960 generic.go:334] "Generic (PLEG): container finished" podID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" containerID="5fc63fba6074f60bae235331fdc246083cbde7e8ad97f7c9af74f179c6bf395b" exitCode=0 Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.238700 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6rmd" event={"ID":"e413a5b2-dcf3-4179-9880-b2cec1a2ef56","Type":"ContainerDied","Data":"5fc63fba6074f60bae235331fdc246083cbde7e8ad97f7c9af74f179c6bf395b"} Dec 01 20:02:33 crc kubenswrapper[4960]: E1201 20:02:33.239422 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.248240 4960 generic.go:334] "Generic (PLEG): container finished" podID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerID="3a2d2dccf519aed23563979e8960e3fb65053dc0fba6f27500aaf3afa6b41645" exitCode=0 Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.248317 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerDied","Data":"3a2d2dccf519aed23563979e8960e3fb65053dc0fba6f27500aaf3afa6b41645"} Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.518654 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.628744 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.628801 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.628916 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.628990 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.629032 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.629090 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6prml\" (UniqueName: \"kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml\") pod \"407668ff-7bd5-44c9-ae04-b479c81d45ce\" (UID: \"407668ff-7bd5-44c9-ae04-b479c81d45ce\") " Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.639054 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml" (OuterVolumeSpecName: "kube-api-access-6prml") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "kube-api-access-6prml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.710027 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.725300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config" (OuterVolumeSpecName: "config") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.731960 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.732079 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6prml\" (UniqueName: \"kubernetes.io/projected/407668ff-7bd5-44c9-ae04-b479c81d45ce-kube-api-access-6prml\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.732189 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.732656 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.750459 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.763687 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "407668ff-7bd5-44c9-ae04-b479c81d45ce" (UID: "407668ff-7bd5-44c9-ae04-b479c81d45ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.833758 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.833795 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.833806 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/407668ff-7bd5-44c9-ae04-b479c81d45ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:33 crc kubenswrapper[4960]: I1201 20:02:33.950430 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.137726 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle\") pod \"80b90b04-9bf0-4bab-a581-4691ef53713e\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.137927 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs\") pod \"80b90b04-9bf0-4bab-a581-4691ef53713e\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.138013 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts\") pod \"80b90b04-9bf0-4bab-a581-4691ef53713e\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.138035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgg4s\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s\") pod \"80b90b04-9bf0-4bab-a581-4691ef53713e\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.138080 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data\") pod \"80b90b04-9bf0-4bab-a581-4691ef53713e\" (UID: \"80b90b04-9bf0-4bab-a581-4691ef53713e\") " Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.143395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s" (OuterVolumeSpecName: "kube-api-access-vgg4s") pod "80b90b04-9bf0-4bab-a581-4691ef53713e" (UID: "80b90b04-9bf0-4bab-a581-4691ef53713e"). InnerVolumeSpecName "kube-api-access-vgg4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.144173 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts" (OuterVolumeSpecName: "scripts") pod "80b90b04-9bf0-4bab-a581-4691ef53713e" (UID: "80b90b04-9bf0-4bab-a581-4691ef53713e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.145514 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs" (OuterVolumeSpecName: "certs") pod "80b90b04-9bf0-4bab-a581-4691ef53713e" (UID: "80b90b04-9bf0-4bab-a581-4691ef53713e"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.169971 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data" (OuterVolumeSpecName: "config-data") pod "80b90b04-9bf0-4bab-a581-4691ef53713e" (UID: "80b90b04-9bf0-4bab-a581-4691ef53713e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.177226 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "80b90b04-9bf0-4bab-a581-4691ef53713e" (UID: "80b90b04-9bf0-4bab-a581-4691ef53713e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.240927 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.240956 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.240966 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgg4s\" (UniqueName: \"kubernetes.io/projected/80b90b04-9bf0-4bab-a581-4691ef53713e-kube-api-access-vgg4s\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.240977 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.240988 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80b90b04-9bf0-4bab-a581-4691ef53713e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.290744 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d89589d9d-9dglh" event={"ID":"9c816e6d-aa4e-46cc-8494-39d3f334a028","Type":"ContainerStarted","Data":"8f1716ecea77f7d9427e47dc59fed7450d830472c1bbf274d618d884d8720e69"} Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.291249 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.291357 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.294972 4960 generic.go:334] "Generic (PLEG): container finished" podID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerID="a5d6130b85a0bd978c56b6b903e0e903220cba1cc09016b3f744f6a4846fddb8" exitCode=0 Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.298044 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerDied","Data":"a5d6130b85a0bd978c56b6b903e0e903220cba1cc09016b3f744f6a4846fddb8"} Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.323489 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-d89589d9d-9dglh" podStartSLOduration=9.323472055 podStartE2EDuration="9.323472055s" podCreationTimestamp="2025-12-01 20:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:34.309539309 +0000 UTC m=+1389.597030978" watchObservedRunningTime="2025-12-01 20:02:34.323472055 +0000 UTC m=+1389.610963724" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.324527 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-8vlfj"] Dec 01 20:02:34 crc kubenswrapper[4960]: E1201 20:02:34.325009 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" containerName="cloudkitty-db-sync" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.325022 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" containerName="cloudkitty-db-sync" Dec 01 20:02:34 crc kubenswrapper[4960]: E1201 20:02:34.325034 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="init" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.325040 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="init" Dec 01 20:02:34 crc kubenswrapper[4960]: E1201 20:02:34.325051 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="dnsmasq-dns" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.325057 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="dnsmasq-dns" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.325337 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" containerName="dnsmasq-dns" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.325356 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" containerName="cloudkitty-db-sync" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.326174 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.335736 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-zdrn9" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.335736 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-zdrn9" event={"ID":"80b90b04-9bf0-4bab-a581-4691ef53713e","Type":"ContainerDied","Data":"faec59cfbf7a7299d1a6802b84ee439c70216a7129d8d046fdc5a0c674b141db"} Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.337661 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faec59cfbf7a7299d1a6802b84ee439c70216a7129d8d046fdc5a0c674b141db" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.347101 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" event={"ID":"407668ff-7bd5-44c9-ae04-b479c81d45ce","Type":"ContainerDied","Data":"990f44e04d54ca2e3300f55059ffbedf6914cf189ce6e5fcc1e2e3a8d18a8dc9"} Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.347323 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-5gct5" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.352229 4960 scope.go:117] "RemoveContainer" containerID="3a2d2dccf519aed23563979e8960e3fb65053dc0fba6f27500aaf3afa6b41645" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.369250 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerStarted","Data":"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78"} Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.369489 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="ceilometer-notification-agent" containerID="cri-o://9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981" gracePeriod=30 Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.369587 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.369630 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="proxy-httpd" containerID="cri-o://954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78" gracePeriod=30 Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.392839 4960 scope.go:117] "RemoveContainer" containerID="c167f048da7a7a034940dab8f14b88399c5feffdcdc95d479ed5f29d50435ffd" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.401454 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8vlfj"] Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.446507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z64l4\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.446548 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.446685 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.446711 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.446767 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.460239 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.481393 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-5gct5"] Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.548861 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.549268 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.549337 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.549400 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z64l4\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.549433 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.565127 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.567487 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.568387 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.568911 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.603560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z64l4\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4\") pod \"cloudkitty-storageinit-8vlfj\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:34 crc kubenswrapper[4960]: I1201 20:02:34.675021 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.013461 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.023496 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.179752 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkqm9\" (UniqueName: \"kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.179811 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.179872 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.179973 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.179999 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.180090 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id\") pod \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\" (UID: \"e413a5b2-dcf3-4179-9880-b2cec1a2ef56\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.181870 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.189148 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.191260 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts" (OuterVolumeSpecName: "scripts") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.197556 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9" (OuterVolumeSpecName: "kube-api-access-zkqm9") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "kube-api-access-zkqm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.221359 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.250394 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.275906 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-8vlfj"] Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.285707 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.285744 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkqm9\" (UniqueName: \"kubernetes.io/projected/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-kube-api-access-zkqm9\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.285757 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.285769 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.285779 4960 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.300953 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data" (OuterVolumeSpecName: "config-data") pod "e413a5b2-dcf3-4179-9880-b2cec1a2ef56" (UID: "e413a5b2-dcf3-4179-9880-b2cec1a2ef56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: W1201 20:02:35.327530 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62a5e00a_99c2_4caa_a683_243167f19130.slice/crio-b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234 WatchSource:0}: Error finding container b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234: Status 404 returned error can't find the container with id b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234 Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.340145 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="407668ff-7bd5-44c9-ae04-b479c81d45ce" path="/var/lib/kubelet/pods/407668ff-7bd5-44c9-ae04-b479c81d45ce/volumes" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.390211 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e413a5b2-dcf3-4179-9880-b2cec1a2ef56-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.437611 4960 generic.go:334] "Generic (PLEG): container finished" podID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerID="bf9b914186246729595c9c68869be37b28c1b46159d8474fe3467995257da219" exitCode=0 Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.437683 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerDied","Data":"bf9b914186246729595c9c68869be37b28c1b46159d8474fe3467995257da219"} Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.470223 4960 generic.go:334] "Generic (PLEG): container finished" podID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerID="954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78" exitCode=0 Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.470304 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerDied","Data":"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78"} Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.481962 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6rmd" event={"ID":"e413a5b2-dcf3-4179-9880-b2cec1a2ef56","Type":"ContainerDied","Data":"87780ef164b901da83309a5993da3e334b0c717350640ef9f5375c7f045a6a38"} Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.482002 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87780ef164b901da83309a5993da3e334b0c717350640ef9f5375c7f045a6a38" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.482072 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6rmd" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.486435 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8vlfj" event={"ID":"62a5e00a-99c2-4caa-a683-243167f19130","Type":"ContainerStarted","Data":"b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234"} Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.550159 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:35 crc kubenswrapper[4960]: E1201 20:02:35.550594 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" containerName="cinder-db-sync" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.550605 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" containerName="cinder-db-sync" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.550806 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" containerName="cinder-db-sync" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.552180 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.561485 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kp5px" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.561722 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.561865 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.563443 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.563971 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.572336 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.661179 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:35 crc kubenswrapper[4960]: E1201 20:02:35.661684 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-api" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.661700 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-api" Dec 01 20:02:35 crc kubenswrapper[4960]: E1201 20:02:35.661721 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-httpd" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.661727 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-httpd" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.661914 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-httpd" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.661931 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" containerName="neutron-api" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.667391 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695299 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle\") pod \"b99c86cc-706e-499a-8b65-7a25d754ac9a\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695386 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config\") pod \"b99c86cc-706e-499a-8b65-7a25d754ac9a\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695434 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config\") pod \"b99c86cc-706e-499a-8b65-7a25d754ac9a\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695478 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h9dv\" (UniqueName: \"kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv\") pod \"b99c86cc-706e-499a-8b65-7a25d754ac9a\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695642 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs\") pod \"b99c86cc-706e-499a-8b65-7a25d754ac9a\" (UID: \"b99c86cc-706e-499a-8b65-7a25d754ac9a\") " Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695885 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695972 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.695994 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.696015 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.696038 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6gpk\" (UniqueName: \"kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.696121 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.712734 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv" (OuterVolumeSpecName: "kube-api-access-6h9dv") pod "b99c86cc-706e-499a-8b65-7a25d754ac9a" (UID: "b99c86cc-706e-499a-8b65-7a25d754ac9a"). InnerVolumeSpecName "kube-api-access-6h9dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.723309 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b99c86cc-706e-499a-8b65-7a25d754ac9a" (UID: "b99c86cc-706e-499a-8b65-7a25d754ac9a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.731222 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.802465 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.802843 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgnfv\" (UniqueName: \"kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.802898 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.802971 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803019 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803043 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6gpk\" (UniqueName: \"kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803158 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803195 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803231 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803237 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config" (OuterVolumeSpecName: "config") pod "b99c86cc-706e-499a-8b65-7a25d754ac9a" (UID: "b99c86cc-706e-499a-8b65-7a25d754ac9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803260 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803361 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803377 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.803442 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h9dv\" (UniqueName: \"kubernetes.io/projected/b99c86cc-706e-499a-8b65-7a25d754ac9a-kube-api-access-6h9dv\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.804806 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.812252 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.812821 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.816042 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.820217 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.832918 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6gpk\" (UniqueName: \"kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk\") pod \"cinder-scheduler-0\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.833293 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b99c86cc-706e-499a-8b65-7a25d754ac9a" (UID: "b99c86cc-706e-499a-8b65-7a25d754ac9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.845931 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b99c86cc-706e-499a-8b65-7a25d754ac9a" (UID: "b99c86cc-706e-499a-8b65-7a25d754ac9a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.871643 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.873312 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.878470 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.913048 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916285 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916429 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgnfv\" (UniqueName: \"kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916612 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916729 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916779 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916877 4960 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.916888 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b99c86cc-706e-499a-8b65-7a25d754ac9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.917727 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.918247 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.920150 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.920856 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.921449 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.948196 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgnfv\" (UniqueName: \"kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv\") pod \"dnsmasq-dns-5c9776ccc5-2jmgw\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:35 crc kubenswrapper[4960]: I1201 20:02:35.955738 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.002871 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018159 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018367 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018451 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz726\" (UniqueName: \"kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018770 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.018884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.122620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.122939 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.122978 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz726\" (UniqueName: \"kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.123020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.123055 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.123075 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.123105 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.127009 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.127025 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.127008 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.131509 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.138694 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.139772 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.146637 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz726\" (UniqueName: \"kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726\") pod \"cinder-api-0\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.157921 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.226684 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.226721 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.227512 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.227590 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.227714 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.227894 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.227993 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.228057 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2272f\" (UniqueName: \"kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.228085 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts\") pod \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\" (UID: \"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf\") " Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.228988 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.229003 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.232136 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts" (OuterVolumeSpecName: "scripts") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.241690 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f" (OuterVolumeSpecName: "kube-api-access-2272f") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "kube-api-access-2272f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.242104 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.292354 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.331613 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.331651 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.331664 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2272f\" (UniqueName: \"kubernetes.io/projected/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-kube-api-access-2272f\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.331673 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.333689 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data" (OuterVolumeSpecName: "config-data") pod "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" (UID: "e7f05fc9-a84c-4ae2-97df-439c5fccf1cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.366452 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.434089 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.526508 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8vlfj" event={"ID":"62a5e00a-99c2-4caa-a683-243167f19130","Type":"ContainerStarted","Data":"2a8938c6209003b731a5d76aa0618ea9d20d4ad73dfb361c0d2f7022bb22c571"} Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.529354 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.531990 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dd6798df6-k67nk" event={"ID":"b99c86cc-706e-499a-8b65-7a25d754ac9a","Type":"ContainerDied","Data":"b44771492dc9e8c4bbf00cdbb631f03cbf065ccc4d7ffbfa17a1f0dcc42bce88"} Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.532045 4960 scope.go:117] "RemoveContainer" containerID="a5d6130b85a0bd978c56b6b903e0e903220cba1cc09016b3f744f6a4846fddb8" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.532238 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dd6798df6-k67nk" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.548031 4960 generic.go:334] "Generic (PLEG): container finished" podID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerID="9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981" exitCode=0 Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.548078 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerDied","Data":"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981"} Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.548104 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e7f05fc9-a84c-4ae2-97df-439c5fccf1cf","Type":"ContainerDied","Data":"6b59ddb4bb1454102201333953de118de892cf1fbc4816a5d08c80c68f5f322b"} Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.548245 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.552245 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-8vlfj" podStartSLOduration=2.552230724 podStartE2EDuration="2.552230724s" podCreationTimestamp="2025-12-01 20:02:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:36.547405829 +0000 UTC m=+1391.834897508" watchObservedRunningTime="2025-12-01 20:02:36.552230724 +0000 UTC m=+1391.839722393" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.618930 4960 scope.go:117] "RemoveContainer" containerID="bf9b914186246729595c9c68869be37b28c1b46159d8474fe3467995257da219" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.640007 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.695368 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dd6798df6-k67nk"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.724311 4960 scope.go:117] "RemoveContainer" containerID="954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78" Dec 01 20:02:36 crc kubenswrapper[4960]: W1201 20:02:36.724318 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf82c85cd_da8d_42e3_ab3c_4a0b12620939.slice/crio-e20403e2b885b81c33114e5884ed78dd1a243638c531e5460539e1067d031ea5 WatchSource:0}: Error finding container e20403e2b885b81c33114e5884ed78dd1a243638c531e5460539e1067d031ea5: Status 404 returned error can't find the container with id e20403e2b885b81c33114e5884ed78dd1a243638c531e5460539e1067d031ea5 Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.743096 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.751011 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.769981 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:02:36 crc kubenswrapper[4960]: E1201 20:02:36.792062 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="ceilometer-notification-agent" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.792406 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="ceilometer-notification-agent" Dec 01 20:02:36 crc kubenswrapper[4960]: E1201 20:02:36.792522 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="proxy-httpd" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.792602 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="proxy-httpd" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.792970 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="ceilometer-notification-agent" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.793070 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" containerName="proxy-httpd" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.795608 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.810902 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.799263 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.816037 4960 scope.go:117] "RemoveContainer" containerID="9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.838599 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:02:36 crc kubenswrapper[4960]: I1201 20:02:36.840418 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015342 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015399 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015423 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkx5j\" (UniqueName: \"kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015484 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015509 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015531 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.015547 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.028540 4960 scope.go:117] "RemoveContainer" containerID="954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78" Dec 01 20:02:37 crc kubenswrapper[4960]: E1201 20:02:37.030284 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78\": container with ID starting with 954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78 not found: ID does not exist" containerID="954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.030328 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78"} err="failed to get container status \"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78\": rpc error: code = NotFound desc = could not find container \"954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78\": container with ID starting with 954b8826c86ce3d45b03133b440367d80b93f64f85ad85afabb05baff7140a78 not found: ID does not exist" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.030352 4960 scope.go:117] "RemoveContainer" containerID="9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981" Dec 01 20:02:37 crc kubenswrapper[4960]: E1201 20:02:37.052286 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981\": container with ID starting with 9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981 not found: ID does not exist" containerID="9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.052327 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981"} err="failed to get container status \"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981\": rpc error: code = NotFound desc = could not find container \"9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981\": container with ID starting with 9aab157d9de6a98107db63868b29ab0f7a1548d30c2c578141111ff6af294981 not found: ID does not exist" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.059719 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119454 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119506 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119579 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119632 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119671 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkx5j\" (UniqueName: \"kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119738 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.119771 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.120281 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.120491 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.126635 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.132861 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.142093 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.145902 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.165792 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkx5j\" (UniqueName: \"kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j\") pod \"ceilometer-0\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.342362 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b99c86cc-706e-499a-8b65-7a25d754ac9a" path="/var/lib/kubelet/pods/b99c86cc-706e-499a-8b65-7a25d754ac9a/volumes" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.343301 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f05fc9-a84c-4ae2-97df-439c5fccf1cf" path="/var/lib/kubelet/pods/e7f05fc9-a84c-4ae2-97df-439c5fccf1cf/volumes" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.351303 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.549472 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.614369 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerStarted","Data":"e2ac69a6cf031a148b3a22d6838c0a7b01f20c9399c5974896c72414425f35c9"} Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.645821 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerStarted","Data":"6008b597b9a24161b7860d47cd8add329417ad60003a80c8a8bfd6057d61aea4"} Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.658429 4960 generic.go:334] "Generic (PLEG): container finished" podID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerID="693930fe97947ce706c32ed52ebbf44081a6f60b3de3e14b92adb1b05b22c3a9" exitCode=0 Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.658624 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" event={"ID":"f82c85cd-da8d-42e3-ab3c-4a0b12620939","Type":"ContainerDied","Data":"693930fe97947ce706c32ed52ebbf44081a6f60b3de3e14b92adb1b05b22c3a9"} Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.659196 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" event={"ID":"f82c85cd-da8d-42e3-ab3c-4a0b12620939","Type":"ContainerStarted","Data":"e20403e2b885b81c33114e5884ed78dd1a243638c531e5460539e1067d031ea5"} Dec 01 20:02:37 crc kubenswrapper[4960]: I1201 20:02:37.932857 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.726099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerStarted","Data":"5fdbe480a5d2707c8943a01ec1e91a1db93049ba8e23331ebfb06202ec12c02f"} Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.729335 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerStarted","Data":"761bbe82eadb796bf9edff9cf73cebfa4391ca054e7b84c10cd9ccada574bde3"} Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.731977 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerStarted","Data":"79431ae1511cc5a17e93fb95f001b2c60d759de6a2718dd3322abbf754a0eaae"} Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.746975 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" event={"ID":"f82c85cd-da8d-42e3-ab3c-4a0b12620939","Type":"ContainerStarted","Data":"3ea77733352468dcf8cf6e81bcfc121b5672cbf6e736b4921546a3f7b182b1c5"} Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.747409 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:38 crc kubenswrapper[4960]: I1201 20:02:38.782818 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" podStartSLOduration=3.78280206 podStartE2EDuration="3.78280206s" podCreationTimestamp="2025-12-01 20:02:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:38.770346831 +0000 UTC m=+1394.057838500" watchObservedRunningTime="2025-12-01 20:02:38.78280206 +0000 UTC m=+1394.070293729" Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.054851 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.756525 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerStarted","Data":"14ceacb722cd22161637bdc4338bf3779f7a72183e4a239b6dc813e685c8a6cc"} Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.758237 4960 generic.go:334] "Generic (PLEG): container finished" podID="62a5e00a-99c2-4caa-a683-243167f19130" containerID="2a8938c6209003b731a5d76aa0618ea9d20d4ad73dfb361c0d2f7022bb22c571" exitCode=0 Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.758293 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8vlfj" event={"ID":"62a5e00a-99c2-4caa-a683-243167f19130","Type":"ContainerDied","Data":"2a8938c6209003b731a5d76aa0618ea9d20d4ad73dfb361c0d2f7022bb22c571"} Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.759840 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerStarted","Data":"56a9a27a582bb6de3091a6c2ec883fc77b2ea4f689f5ace6a4ccc4b8f8df2c95"} Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.760246 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api-log" containerID="cri-o://5fdbe480a5d2707c8943a01ec1e91a1db93049ba8e23331ebfb06202ec12c02f" gracePeriod=30 Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.760300 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api" containerID="cri-o://56a9a27a582bb6de3091a6c2ec883fc77b2ea4f689f5ace6a4ccc4b8f8df2c95" gracePeriod=30 Dec 01 20:02:39 crc kubenswrapper[4960]: I1201 20:02:39.830334 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.830314105 podStartE2EDuration="4.830314105s" podCreationTimestamp="2025-12-01 20:02:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:39.81861483 +0000 UTC m=+1395.106106489" watchObservedRunningTime="2025-12-01 20:02:39.830314105 +0000 UTC m=+1395.117805774" Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.775815 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerStarted","Data":"da8bd8fd669d1d3640e5fc203a3ab3729a8491d957d3b27428ad827bd782a297"} Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.778329 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerStarted","Data":"689ffbbc20f65a6468ccba06177d0ed96525005202a475890c81476bf6b37765"} Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.780831 4960 generic.go:334] "Generic (PLEG): container finished" podID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerID="5fdbe480a5d2707c8943a01ec1e91a1db93049ba8e23331ebfb06202ec12c02f" exitCode=143 Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.780914 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerDied","Data":"5fdbe480a5d2707c8943a01ec1e91a1db93049ba8e23331ebfb06202ec12c02f"} Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.810495 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.688527147 podStartE2EDuration="5.810475094s" podCreationTimestamp="2025-12-01 20:02:35 +0000 UTC" firstStartedPulling="2025-12-01 20:02:36.544487526 +0000 UTC m=+1391.831979195" lastFinishedPulling="2025-12-01 20:02:37.666435473 +0000 UTC m=+1392.953927142" observedRunningTime="2025-12-01 20:02:40.801533238 +0000 UTC m=+1396.089024907" watchObservedRunningTime="2025-12-01 20:02:40.810475094 +0000 UTC m=+1396.097966763" Dec 01 20:02:40 crc kubenswrapper[4960]: I1201 20:02:40.955879 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.263717 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.358855 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs\") pod \"62a5e00a-99c2-4caa-a683-243167f19130\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.359103 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z64l4\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4\") pod \"62a5e00a-99c2-4caa-a683-243167f19130\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.359162 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data\") pod \"62a5e00a-99c2-4caa-a683-243167f19130\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.359199 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts\") pod \"62a5e00a-99c2-4caa-a683-243167f19130\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.359321 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle\") pod \"62a5e00a-99c2-4caa-a683-243167f19130\" (UID: \"62a5e00a-99c2-4caa-a683-243167f19130\") " Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.363858 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs" (OuterVolumeSpecName: "certs") pod "62a5e00a-99c2-4caa-a683-243167f19130" (UID: "62a5e00a-99c2-4caa-a683-243167f19130"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.385165 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts" (OuterVolumeSpecName: "scripts") pod "62a5e00a-99c2-4caa-a683-243167f19130" (UID: "62a5e00a-99c2-4caa-a683-243167f19130"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.387812 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4" (OuterVolumeSpecName: "kube-api-access-z64l4") pod "62a5e00a-99c2-4caa-a683-243167f19130" (UID: "62a5e00a-99c2-4caa-a683-243167f19130"). InnerVolumeSpecName "kube-api-access-z64l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.395241 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62a5e00a-99c2-4caa-a683-243167f19130" (UID: "62a5e00a-99c2-4caa-a683-243167f19130"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.401958 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data" (OuterVolumeSpecName: "config-data") pod "62a5e00a-99c2-4caa-a683-243167f19130" (UID: "62a5e00a-99c2-4caa-a683-243167f19130"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.417626 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.462134 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.462364 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z64l4\" (UniqueName: \"kubernetes.io/projected/62a5e00a-99c2-4caa-a683-243167f19130-kube-api-access-z64l4\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.462440 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.462499 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.462560 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62a5e00a-99c2-4caa-a683-243167f19130-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.796664 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerStarted","Data":"d760d4e9de8c02a457eb77823cd01dc91c251788bfc958df8ea38e97db79399b"} Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.799585 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-8vlfj" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.799595 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-8vlfj" event={"ID":"62a5e00a-99c2-4caa-a683-243167f19130","Type":"ContainerDied","Data":"b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234"} Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.799647 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7512471454cb6ed025d1bf63731853f4b5201d55255cac514a1e647ccf11234" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.984072 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:41 crc kubenswrapper[4960]: E1201 20:02:41.984615 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a5e00a-99c2-4caa-a683-243167f19130" containerName="cloudkitty-storageinit" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.984632 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a5e00a-99c2-4caa-a683-243167f19130" containerName="cloudkitty-storageinit" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.984846 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a5e00a-99c2-4caa-a683-243167f19130" containerName="cloudkitty-storageinit" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.985779 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.991687 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-scripts" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.996260 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-cloudkitty-dockercfg-ndz6g" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.996503 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-client-internal" Dec 01 20:02:41 crc kubenswrapper[4960]: I1201 20:02:41.997056 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.004501 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.003919 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-config-data" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080414 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080528 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080552 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080575 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080635 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.080686 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qm4r\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.204995 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.205169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.205195 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.205218 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.205314 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.205384 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qm4r\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.226966 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.237740 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qm4r\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.238132 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.241355 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.246288 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="dnsmasq-dns" containerID="cri-o://3ea77733352468dcf8cf6e81bcfc121b5672cbf6e736b4921546a3f7b182b1c5" gracePeriod=10 Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.296845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.297952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.298353 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs\") pod \"cloudkitty-proc-0\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.315537 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.317655 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.337988 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.348224 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422557 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422655 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422712 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422729 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbcxc\" (UniqueName: \"kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.422797 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.495337 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.498531 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.505713 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.518037 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.524682 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525488 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525591 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525612 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbcxc\" (UniqueName: \"kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525673 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.525740 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.526297 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.526800 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.528430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.533646 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.590510 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbcxc\" (UniqueName: \"kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc\") pod \"dnsmasq-dns-67bdc55879-5cl7g\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628376 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628419 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzfk2\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628458 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628483 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628576 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.628604 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.667991 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.731672 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.731781 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732093 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732186 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732298 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732330 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732428 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzfk2\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.732449 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.737754 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.738238 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.738950 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.742137 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.754874 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.760746 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzfk2\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2\") pod \"cloudkitty-api-0\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.822241 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.824389 4960 generic.go:334] "Generic (PLEG): container finished" podID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerID="3ea77733352468dcf8cf6e81bcfc121b5672cbf6e736b4921546a3f7b182b1c5" exitCode=0 Dec 01 20:02:42 crc kubenswrapper[4960]: I1201 20:02:42.825388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" event={"ID":"f82c85cd-da8d-42e3-ab3c-4a0b12620939","Type":"ContainerDied","Data":"3ea77733352468dcf8cf6e81bcfc121b5672cbf6e736b4921546a3f7b182b1c5"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.057724 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.175990 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d89589d9d-9dglh" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.292324 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.292545 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6877cc4654-npk8m" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api-log" containerID="cri-o://1120540dd51135b7e824078ba3e461c05d58dc1f0991bb8342834740740e88c5" gracePeriod=30 Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.292678 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6877cc4654-npk8m" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api" containerID="cri-o://2def03205cf6fc6af7659316ce5aeee330d7d869f904b59a378ee0fa76edec05" gracePeriod=30 Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.504578 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.527646 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:43 crc kubenswrapper[4960]: W1201 20:02:43.528243 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18497e75_1fc4_4324_ab61_b067d19c1ce0.slice/crio-0f2543b2077d3ab6df6492fde1518e407dd1246d1d90d7831a9e511a878b1fa6 WatchSource:0}: Error finding container 0f2543b2077d3ab6df6492fde1518e407dd1246d1d90d7831a9e511a878b1fa6: Status 404 returned error can't find the container with id 0f2543b2077d3ab6df6492fde1518e407dd1246d1d90d7831a9e511a878b1fa6 Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563354 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563457 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563489 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563542 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgnfv\" (UniqueName: \"kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563583 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.563637 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.578915 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv" (OuterVolumeSpecName: "kube-api-access-xgnfv") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "kube-api-access-xgnfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.645849 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config" (OuterVolumeSpecName: "config") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.664780 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.666743 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.667558 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") pod \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\" (UID: \"f82c85cd-da8d-42e3-ab3c-4a0b12620939\") " Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.670659 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.670683 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.670693 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgnfv\" (UniqueName: \"kubernetes.io/projected/f82c85cd-da8d-42e3-ab3c-4a0b12620939-kube-api-access-xgnfv\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: W1201 20:02:43.670800 4960 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f82c85cd-da8d-42e3-ab3c-4a0b12620939/volumes/kubernetes.io~configmap/ovsdbserver-sb Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.670813 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.681770 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.710806 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f82c85cd-da8d-42e3-ab3c-4a0b12620939" (UID: "f82c85cd-da8d-42e3-ab3c-4a0b12620939"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.726507 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.773523 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.774011 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.774026 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f82c85cd-da8d-42e3-ab3c-4a0b12620939-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.841173 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerStarted","Data":"82cc2058ad2c0b1eab96b25493b6333a7eccb587f45be071c55c03fe30bb48cf"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.844047 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"b709c481-2fd8-4b78-b39c-7be73c764fe8","Type":"ContainerStarted","Data":"69faf6b590f580088c64356d0c09002ec2f74779034845616be7232299e82489"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.850304 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerStarted","Data":"f25fee3ed7a2dbc06c02b73be487d6baed181afa283327e83755925d4c1968ff"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.851655 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.870910 4960 generic.go:334] "Generic (PLEG): container finished" podID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerID="1120540dd51135b7e824078ba3e461c05d58dc1f0991bb8342834740740e88c5" exitCode=143 Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.871030 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerDied","Data":"1120540dd51135b7e824078ba3e461c05d58dc1f0991bb8342834740740e88c5"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.877620 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9139700189999997 podStartE2EDuration="7.877601733s" podCreationTimestamp="2025-12-01 20:02:36 +0000 UTC" firstStartedPulling="2025-12-01 20:02:37.948157229 +0000 UTC m=+1393.235648898" lastFinishedPulling="2025-12-01 20:02:42.911788933 +0000 UTC m=+1398.199280612" observedRunningTime="2025-12-01 20:02:43.872741378 +0000 UTC m=+1399.160233047" watchObservedRunningTime="2025-12-01 20:02:43.877601733 +0000 UTC m=+1399.165093402" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.878014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerStarted","Data":"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.878055 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerStarted","Data":"0f2543b2077d3ab6df6492fde1518e407dd1246d1d90d7831a9e511a878b1fa6"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.890421 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" event={"ID":"f82c85cd-da8d-42e3-ab3c-4a0b12620939","Type":"ContainerDied","Data":"e20403e2b885b81c33114e5884ed78dd1a243638c531e5460539e1067d031ea5"} Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.890471 4960 scope.go:117] "RemoveContainer" containerID="3ea77733352468dcf8cf6e81bcfc121b5672cbf6e736b4921546a3f7b182b1c5" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.890525 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-2jmgw" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.929813 4960 scope.go:117] "RemoveContainer" containerID="693930fe97947ce706c32ed52ebbf44081a6f60b3de3e14b92adb1b05b22c3a9" Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.949075 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:43 crc kubenswrapper[4960]: I1201 20:02:43.961922 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-2jmgw"] Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.924453 4960 generic.go:334] "Generic (PLEG): container finished" podID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerID="1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b" exitCode=0 Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.924828 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerDied","Data":"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b"} Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.924865 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.924875 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerStarted","Data":"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25"} Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.931731 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerStarted","Data":"7834d7b50b4c1ce277602b9a23a8073962fa2ac7d738f08177af86520cb56be6"} Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.931767 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.931777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerStarted","Data":"02919cffee6a9d3fa016487ac95bf59a65e31512e2bf4e3ccbd8b2d3f3d8cda5"} Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.981551 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" podStartSLOduration=2.981533793 podStartE2EDuration="2.981533793s" podCreationTimestamp="2025-12-01 20:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:44.957181434 +0000 UTC m=+1400.244673103" watchObservedRunningTime="2025-12-01 20:02:44.981533793 +0000 UTC m=+1400.269025462" Dec 01 20:02:44 crc kubenswrapper[4960]: I1201 20:02:44.988212 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.988196846 podStartE2EDuration="2.988196846s" podCreationTimestamp="2025-12-01 20:02:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:44.98049852 +0000 UTC m=+1400.267990189" watchObservedRunningTime="2025-12-01 20:02:44.988196846 +0000 UTC m=+1400.275688505" Dec 01 20:02:45 crc kubenswrapper[4960]: I1201 20:02:45.345097 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" path="/var/lib/kubelet/pods/f82c85cd-da8d-42e3-ab3c-4a0b12620939/volumes" Dec 01 20:02:45 crc kubenswrapper[4960]: I1201 20:02:45.455042 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:45 crc kubenswrapper[4960]: I1201 20:02:45.942302 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"b709c481-2fd8-4b78-b39c-7be73c764fe8","Type":"ContainerStarted","Data":"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09"} Dec 01 20:02:45 crc kubenswrapper[4960]: I1201 20:02:45.969431 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=3.24402901 podStartE2EDuration="4.969410729s" podCreationTimestamp="2025-12-01 20:02:41 +0000 UTC" firstStartedPulling="2025-12-01 20:02:43.062583329 +0000 UTC m=+1398.350074998" lastFinishedPulling="2025-12-01 20:02:44.787965048 +0000 UTC m=+1400.075456717" observedRunningTime="2025-12-01 20:02:45.962788617 +0000 UTC m=+1401.250280306" watchObservedRunningTime="2025-12-01 20:02:45.969410729 +0000 UTC m=+1401.256902408" Dec 01 20:02:45 crc kubenswrapper[4960]: I1201 20:02:45.993261 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:46 crc kubenswrapper[4960]: I1201 20:02:46.213885 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 20:02:46 crc kubenswrapper[4960]: I1201 20:02:46.261453 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.062722 4960 generic.go:334] "Generic (PLEG): container finished" podID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerID="2def03205cf6fc6af7659316ce5aeee330d7d869f904b59a378ee0fa76edec05" exitCode=0 Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.063277 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api-log" containerID="cri-o://02919cffee6a9d3fa016487ac95bf59a65e31512e2bf4e3ccbd8b2d3f3d8cda5" gracePeriod=30 Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.063606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerDied","Data":"2def03205cf6fc6af7659316ce5aeee330d7d869f904b59a378ee0fa76edec05"} Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.064432 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="cinder-scheduler" containerID="cri-o://761bbe82eadb796bf9edff9cf73cebfa4391ca054e7b84c10cd9ccada574bde3" gracePeriod=30 Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.064717 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api" containerID="cri-o://7834d7b50b4c1ce277602b9a23a8073962fa2ac7d738f08177af86520cb56be6" gracePeriod=30 Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.064775 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="probe" containerID="cri-o://da8bd8fd669d1d3640e5fc203a3ab3729a8491d957d3b27428ad827bd782a297" gracePeriod=30 Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.447556 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.596843 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8l65\" (UniqueName: \"kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65\") pod \"5775a86b-8d53-4b08-a420-196487d9e2bf\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.596904 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom\") pod \"5775a86b-8d53-4b08-a420-196487d9e2bf\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.596934 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle\") pod \"5775a86b-8d53-4b08-a420-196487d9e2bf\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.597180 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data\") pod \"5775a86b-8d53-4b08-a420-196487d9e2bf\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.597216 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs\") pod \"5775a86b-8d53-4b08-a420-196487d9e2bf\" (UID: \"5775a86b-8d53-4b08-a420-196487d9e2bf\") " Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.598040 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs" (OuterVolumeSpecName: "logs") pod "5775a86b-8d53-4b08-a420-196487d9e2bf" (UID: "5775a86b-8d53-4b08-a420-196487d9e2bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.614331 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65" (OuterVolumeSpecName: "kube-api-access-h8l65") pod "5775a86b-8d53-4b08-a420-196487d9e2bf" (UID: "5775a86b-8d53-4b08-a420-196487d9e2bf"). InnerVolumeSpecName "kube-api-access-h8l65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.629310 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5775a86b-8d53-4b08-a420-196487d9e2bf" (UID: "5775a86b-8d53-4b08-a420-196487d9e2bf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.676337 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5775a86b-8d53-4b08-a420-196487d9e2bf" (UID: "5775a86b-8d53-4b08-a420-196487d9e2bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.700754 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5775a86b-8d53-4b08-a420-196487d9e2bf-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.700794 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8l65\" (UniqueName: \"kubernetes.io/projected/5775a86b-8d53-4b08-a420-196487d9e2bf-kube-api-access-h8l65\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.700805 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.700814 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.740262 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data" (OuterVolumeSpecName: "config-data") pod "5775a86b-8d53-4b08-a420-196487d9e2bf" (UID: "5775a86b-8d53-4b08-a420-196487d9e2bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:47 crc kubenswrapper[4960]: I1201 20:02:47.803786 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5775a86b-8d53-4b08-a420-196487d9e2bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.087270 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6877cc4654-npk8m" event={"ID":"5775a86b-8d53-4b08-a420-196487d9e2bf","Type":"ContainerDied","Data":"04dba8922d50276f88c48b40f248109476dcd396863bfa309c4041a211acf1dd"} Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.087573 4960 scope.go:117] "RemoveContainer" containerID="2def03205cf6fc6af7659316ce5aeee330d7d869f904b59a378ee0fa76edec05" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.087686 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6877cc4654-npk8m" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.098308 4960 generic.go:334] "Generic (PLEG): container finished" podID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerID="7834d7b50b4c1ce277602b9a23a8073962fa2ac7d738f08177af86520cb56be6" exitCode=0 Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.098362 4960 generic.go:334] "Generic (PLEG): container finished" podID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerID="02919cffee6a9d3fa016487ac95bf59a65e31512e2bf4e3ccbd8b2d3f3d8cda5" exitCode=143 Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.098532 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="b709c481-2fd8-4b78-b39c-7be73c764fe8" containerName="cloudkitty-proc" containerID="cri-o://9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09" gracePeriod=30 Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.099040 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerDied","Data":"7834d7b50b4c1ce277602b9a23a8073962fa2ac7d738f08177af86520cb56be6"} Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.099091 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerDied","Data":"02919cffee6a9d3fa016487ac95bf59a65e31512e2bf4e3ccbd8b2d3f3d8cda5"} Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.182551 4960 scope.go:117] "RemoveContainer" containerID="1120540dd51135b7e824078ba3e461c05d58dc1f0991bb8342834740740e88c5" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.204803 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.220298 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.231641 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6877cc4654-npk8m"] Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.318680 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.318780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.318880 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.318902 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.319015 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzfk2\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.319034 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.319078 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts\") pod \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\" (UID: \"2d55c5b5-86e0-4f7c-9178-178e555b8f44\") " Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.319541 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs" (OuterVolumeSpecName: "logs") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.320091 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d55c5b5-86e0-4f7c-9178-178e555b8f44-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.327600 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs" (OuterVolumeSpecName: "certs") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.327807 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts" (OuterVolumeSpecName: "scripts") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.328378 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.332465 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2" (OuterVolumeSpecName: "kube-api-access-qzfk2") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "kube-api-access-qzfk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.362318 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.365882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data" (OuterVolumeSpecName: "config-data") pod "2d55c5b5-86e0-4f7c-9178-178e555b8f44" (UID: "2d55c5b5-86e0-4f7c-9178-178e555b8f44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422853 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422946 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422955 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422966 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzfk2\" (UniqueName: \"kubernetes.io/projected/2d55c5b5-86e0-4f7c-9178-178e555b8f44-kube-api-access-qzfk2\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422976 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:48 crc kubenswrapper[4960]: I1201 20:02:48.422984 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d55c5b5-86e0-4f7c-9178-178e555b8f44-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.056281 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.108506 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"2d55c5b5-86e0-4f7c-9178-178e555b8f44","Type":"ContainerDied","Data":"82cc2058ad2c0b1eab96b25493b6333a7eccb587f45be071c55c03fe30bb48cf"} Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.108562 4960 scope.go:117] "RemoveContainer" containerID="7834d7b50b4c1ce277602b9a23a8073962fa2ac7d738f08177af86520cb56be6" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.108664 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.117299 4960 generic.go:334] "Generic (PLEG): container finished" podID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerID="da8bd8fd669d1d3640e5fc203a3ab3729a8491d957d3b27428ad827bd782a297" exitCode=0 Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.117394 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerDied","Data":"da8bd8fd669d1d3640e5fc203a3ab3729a8491d957d3b27428ad827bd782a297"} Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.133398 4960 scope.go:117] "RemoveContainer" containerID="02919cffee6a9d3fa016487ac95bf59a65e31512e2bf4e3ccbd8b2d3f3d8cda5" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.154775 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.185749 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199103 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199622 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199635 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199647 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199653 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api" Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199668 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199674 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api" Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199700 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="init" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199706 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="init" Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199717 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199722 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: E1201 20:02:49.199734 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="dnsmasq-dns" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199740 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="dnsmasq-dns" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199927 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199939 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f82c85cd-da8d-42e3-ab3c-4a0b12620939" containerName="dnsmasq-dns" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199953 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api-log" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199966 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" containerName="cloudkitty-api" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.199977 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" containerName="barbican-api" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.201137 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.203737 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.204607 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.204775 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.215795 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.338327 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d55c5b5-86e0-4f7c-9178-178e555b8f44" path="/var/lib/kubelet/pods/2d55c5b5-86e0-4f7c-9178-178e555b8f44/volumes" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339391 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5775a86b-8d53-4b08-a420-196487d9e2bf" path="/var/lib/kubelet/pods/5775a86b-8d53-4b08-a420-196487d9e2bf/volumes" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339493 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339529 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339584 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339635 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339732 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xssk8\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.339950 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.340051 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.340135 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.340241 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442560 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442655 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442693 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442738 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442839 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442880 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442918 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.442951 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xssk8\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.444784 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.447945 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.448645 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.448772 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.448857 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.453096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.453573 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.454851 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.463615 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xssk8\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8\") pod \"cloudkitty-api-0\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " pod="openstack/cloudkitty-api-0" Dec 01 20:02:49 crc kubenswrapper[4960]: I1201 20:02:49.524499 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.039884 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:02:50 crc kubenswrapper[4960]: W1201 20:02:50.040160 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd821084c_2f3c_413c_9c98_e78f7706a75b.slice/crio-df74ff1da0635fa806a304ebb7fc894f4c0405c7f4b3cedf87fbe86b729cd877 WatchSource:0}: Error finding container df74ff1da0635fa806a304ebb7fc894f4c0405c7f4b3cedf87fbe86b729cd877: Status 404 returned error can't find the container with id df74ff1da0635fa806a304ebb7fc894f4c0405c7f4b3cedf87fbe86b729cd877 Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.133171 4960 generic.go:334] "Generic (PLEG): container finished" podID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerID="761bbe82eadb796bf9edff9cf73cebfa4391ca054e7b84c10cd9ccada574bde3" exitCode=0 Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.133272 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerDied","Data":"761bbe82eadb796bf9edff9cf73cebfa4391ca054e7b84c10cd9ccada574bde3"} Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.134050 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerStarted","Data":"df74ff1da0635fa806a304ebb7fc894f4c0405c7f4b3cedf87fbe86b729cd877"} Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.624980 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692444 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692707 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692767 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692795 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6gpk\" (UniqueName: \"kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692835 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692894 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.692966 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts\") pod \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\" (UID: \"19df5099-f6b5-4c5e-81f1-18b3abbacc54\") " Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.693535 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/19df5099-f6b5-4c5e-81f1-18b3abbacc54-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.696975 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.697381 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk" (OuterVolumeSpecName: "kube-api-access-p6gpk") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "kube-api-access-p6gpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.697878 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts" (OuterVolumeSpecName: "scripts") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.744431 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.796091 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.796274 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6gpk\" (UniqueName: \"kubernetes.io/projected/19df5099-f6b5-4c5e-81f1-18b3abbacc54-kube-api-access-p6gpk\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.796289 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.796298 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.797231 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data" (OuterVolumeSpecName: "config-data") pod "19df5099-f6b5-4c5e-81f1-18b3abbacc54" (UID: "19df5099-f6b5-4c5e-81f1-18b3abbacc54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:50 crc kubenswrapper[4960]: I1201 20:02:50.898289 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19df5099-f6b5-4c5e-81f1-18b3abbacc54-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.154352 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"19df5099-f6b5-4c5e-81f1-18b3abbacc54","Type":"ContainerDied","Data":"6008b597b9a24161b7860d47cd8add329417ad60003a80c8a8bfd6057d61aea4"} Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.154381 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.154423 4960 scope.go:117] "RemoveContainer" containerID="da8bd8fd669d1d3640e5fc203a3ab3729a8491d957d3b27428ad827bd782a297" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.158916 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerStarted","Data":"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac"} Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.159250 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerStarted","Data":"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a"} Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.160015 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.191243 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.191220907 podStartE2EDuration="2.191220907s" podCreationTimestamp="2025-12-01 20:02:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:51.179094149 +0000 UTC m=+1406.466585838" watchObservedRunningTime="2025-12-01 20:02:51.191220907 +0000 UTC m=+1406.478712576" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.201722 4960 scope.go:117] "RemoveContainer" containerID="761bbe82eadb796bf9edff9cf73cebfa4391ca054e7b84c10cd9ccada574bde3" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.207014 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.225891 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.236257 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:51 crc kubenswrapper[4960]: E1201 20:02:51.236683 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="cinder-scheduler" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.236699 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="cinder-scheduler" Dec 01 20:02:51 crc kubenswrapper[4960]: E1201 20:02:51.236728 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="probe" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.236735 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="probe" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.236930 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="probe" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.236951 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" containerName="cinder-scheduler" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.238094 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.243041 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.253698 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.305679 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.305733 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.305816 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0bee6fd-952d-4a33-83dc-c93625f15979-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.306026 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tg4p\" (UniqueName: \"kubernetes.io/projected/c0bee6fd-952d-4a33-83dc-c93625f15979-kube-api-access-4tg4p\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.306060 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.306148 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.339376 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19df5099-f6b5-4c5e-81f1-18b3abbacc54" path="/var/lib/kubelet/pods/19df5099-f6b5-4c5e-81f1-18b3abbacc54/volumes" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408167 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408226 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0bee6fd-952d-4a33-83dc-c93625f15979-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408339 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tg4p\" (UniqueName: \"kubernetes.io/projected/c0bee6fd-952d-4a33-83dc-c93625f15979-kube-api-access-4tg4p\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408369 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408432 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408479 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.408504 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0bee6fd-952d-4a33-83dc-c93625f15979-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.414013 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.414462 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-scripts\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.415084 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.417027 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0bee6fd-952d-4a33-83dc-c93625f15979-config-data\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.428787 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tg4p\" (UniqueName: \"kubernetes.io/projected/c0bee6fd-952d-4a33-83dc-c93625f15979-kube-api-access-4tg4p\") pod \"cinder-scheduler-0\" (UID: \"c0bee6fd-952d-4a33-83dc-c93625f15979\") " pod="openstack/cinder-scheduler-0" Dec 01 20:02:51 crc kubenswrapper[4960]: I1201 20:02:51.562187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.043316 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 01 20:02:52 crc kubenswrapper[4960]: W1201 20:02:52.047941 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0bee6fd_952d_4a33_83dc_c93625f15979.slice/crio-7878adcedb0393b252f07ec45fa9d17f9f1cd2f6028a74be12749e4a557364d4 WatchSource:0}: Error finding container 7878adcedb0393b252f07ec45fa9d17f9f1cd2f6028a74be12749e4a557364d4: Status 404 returned error can't find the container with id 7878adcedb0393b252f07ec45fa9d17f9f1cd2f6028a74be12749e4a557364d4 Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.175720 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c0bee6fd-952d-4a33-83dc-c93625f15979","Type":"ContainerStarted","Data":"7878adcedb0393b252f07ec45fa9d17f9f1cd2f6028a74be12749e4a557364d4"} Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.670294 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.784951 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.785617 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="dnsmasq-dns" containerID="cri-o://2dc6c104449fbff3d2ac09a2e4fc67f5471e625b02d903621c8545533d465795" gracePeriod=10 Dec 01 20:02:52 crc kubenswrapper[4960]: I1201 20:02:52.866256 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.176:5353: connect: connection refused" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.186428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c0bee6fd-952d-4a33-83dc-c93625f15979","Type":"ContainerStarted","Data":"05fc3a3c4fa5d3a176d0be8a6e3e2b3d5ea534ad9294528b7e2c76d62987b7ae"} Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.190632 4960 generic.go:334] "Generic (PLEG): container finished" podID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerID="2dc6c104449fbff3d2ac09a2e4fc67f5471e625b02d903621c8545533d465795" exitCode=0 Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.190666 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" event={"ID":"453fc2ad-627e-4256-88f8-6c27cd68211e","Type":"ContainerDied","Data":"2dc6c104449fbff3d2ac09a2e4fc67f5471e625b02d903621c8545533d465795"} Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.596318 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.680777 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdrw4\" (UniqueName: \"kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.680845 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.680886 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.681030 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.681051 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.681187 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc\") pod \"453fc2ad-627e-4256-88f8-6c27cd68211e\" (UID: \"453fc2ad-627e-4256-88f8-6c27cd68211e\") " Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.694695 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4" (OuterVolumeSpecName: "kube-api-access-wdrw4") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "kube-api-access-wdrw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.784608 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdrw4\" (UniqueName: \"kubernetes.io/projected/453fc2ad-627e-4256-88f8-6c27cd68211e-kube-api-access-wdrw4\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.901877 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.903194 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.922162 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config" (OuterVolumeSpecName: "config") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.935361 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.936388 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "453fc2ad-627e-4256-88f8-6c27cd68211e" (UID: "453fc2ad-627e-4256-88f8-6c27cd68211e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.985759 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.988988 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.989017 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.989028 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.989038 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:53 crc kubenswrapper[4960]: I1201 20:02:53.989046 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453fc2ad-627e-4256-88f8-6c27cd68211e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.090362 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.090770 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.090887 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qm4r\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.090967 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.091006 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.091080 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom\") pod \"b709c481-2fd8-4b78-b39c-7be73c764fe8\" (UID: \"b709c481-2fd8-4b78-b39c-7be73c764fe8\") " Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.096398 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.096701 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs" (OuterVolumeSpecName: "certs") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.097900 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts" (OuterVolumeSpecName: "scripts") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.100462 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r" (OuterVolumeSpecName: "kube-api-access-4qm4r") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "kube-api-access-4qm4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.127864 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data" (OuterVolumeSpecName: "config-data") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.128407 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b709c481-2fd8-4b78-b39c-7be73c764fe8" (UID: "b709c481-2fd8-4b78-b39c-7be73c764fe8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193651 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193705 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193714 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193731 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qm4r\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-kube-api-access-4qm4r\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193762 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b709c481-2fd8-4b78-b39c-7be73c764fe8-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.193770 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/b709c481-2fd8-4b78-b39c-7be73c764fe8-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.224728 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" event={"ID":"453fc2ad-627e-4256-88f8-6c27cd68211e","Type":"ContainerDied","Data":"3c0b02161d795043b2d643ef24f550ce1d5cf2a886c99a7b342129cee852429c"} Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.224781 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-5xx8l" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.224807 4960 scope.go:117] "RemoveContainer" containerID="2dc6c104449fbff3d2ac09a2e4fc67f5471e625b02d903621c8545533d465795" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.229209 4960 generic.go:334] "Generic (PLEG): container finished" podID="b709c481-2fd8-4b78-b39c-7be73c764fe8" containerID="9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09" exitCode=0 Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.229327 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.229352 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"b709c481-2fd8-4b78-b39c-7be73c764fe8","Type":"ContainerDied","Data":"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09"} Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.229387 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"b709c481-2fd8-4b78-b39c-7be73c764fe8","Type":"ContainerDied","Data":"69faf6b590f580088c64356d0c09002ec2f74779034845616be7232299e82489"} Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.244556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c0bee6fd-952d-4a33-83dc-c93625f15979","Type":"ContainerStarted","Data":"d06b510aeace15528423358815325deee13488a1ac238851ba28eaeba928f0ae"} Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.309227 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.309207084 podStartE2EDuration="3.309207084s" podCreationTimestamp="2025-12-01 20:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:54.284902087 +0000 UTC m=+1409.572393766" watchObservedRunningTime="2025-12-01 20:02:54.309207084 +0000 UTC m=+1409.596698753" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.321880 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.344891 4960 scope.go:117] "RemoveContainer" containerID="d8e61c199e3033b465cc14b5e1515159c3cf03214cebfe1ce0d8f1ca93aa8f28" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.360502 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.374763 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.386624 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-5xx8l"] Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.399582 4960 scope.go:117] "RemoveContainer" containerID="9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.403886 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:54 crc kubenswrapper[4960]: E1201 20:02:54.404395 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="init" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.404416 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="init" Dec 01 20:02:54 crc kubenswrapper[4960]: E1201 20:02:54.404436 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="dnsmasq-dns" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.404448 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="dnsmasq-dns" Dec 01 20:02:54 crc kubenswrapper[4960]: E1201 20:02:54.404493 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b709c481-2fd8-4b78-b39c-7be73c764fe8" containerName="cloudkitty-proc" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.404502 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b709c481-2fd8-4b78-b39c-7be73c764fe8" containerName="cloudkitty-proc" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.404705 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b709c481-2fd8-4b78-b39c-7be73c764fe8" containerName="cloudkitty-proc" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.404741 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" containerName="dnsmasq-dns" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.405538 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.408064 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.419366 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.465282 4960 scope.go:117] "RemoveContainer" containerID="9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09" Dec 01 20:02:54 crc kubenswrapper[4960]: E1201 20:02:54.466091 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09\": container with ID starting with 9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09 not found: ID does not exist" containerID="9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.466141 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09"} err="failed to get container status \"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09\": rpc error: code = NotFound desc = could not find container \"9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09\": container with ID starting with 9e5fea292df1cb0e51fe0c66f70b2796fa7973e1f28be2bf3f6e9fe954e8dc09 not found: ID does not exist" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.500989 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.501047 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.501087 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zds2p\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.501164 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.501200 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.501511 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.531878 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.538982 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-74c7bf679b-vg6f9" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603387 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603482 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603512 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603546 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zds2p\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603578 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.603613 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.614713 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.614842 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.616243 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.616768 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.625027 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.635705 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zds2p\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p\") pod \"cloudkitty-proc-0\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:02:54 crc kubenswrapper[4960]: I1201 20:02:54.764257 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:02:55 crc kubenswrapper[4960]: I1201 20:02:55.092967 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5f8c78b6fc-fdfln" Dec 01 20:02:55 crc kubenswrapper[4960]: W1201 20:02:55.249024 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21d1f29c_ee4b_4009_8017_ee502c3cfab9.slice/crio-7026f6c36da5fc3be046075c10236ac5282cec17eabcea341248872260154650 WatchSource:0}: Error finding container 7026f6c36da5fc3be046075c10236ac5282cec17eabcea341248872260154650: Status 404 returned error can't find the container with id 7026f6c36da5fc3be046075c10236ac5282cec17eabcea341248872260154650 Dec 01 20:02:55 crc kubenswrapper[4960]: I1201 20:02:55.259958 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:02:55 crc kubenswrapper[4960]: I1201 20:02:55.364980 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="453fc2ad-627e-4256-88f8-6c27cd68211e" path="/var/lib/kubelet/pods/453fc2ad-627e-4256-88f8-6c27cd68211e/volumes" Dec 01 20:02:55 crc kubenswrapper[4960]: I1201 20:02:55.365765 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b709c481-2fd8-4b78-b39c-7be73c764fe8" path="/var/lib/kubelet/pods/b709c481-2fd8-4b78-b39c-7be73c764fe8/volumes" Dec 01 20:02:56 crc kubenswrapper[4960]: I1201 20:02:56.281961 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"21d1f29c-ee4b-4009-8017-ee502c3cfab9","Type":"ContainerStarted","Data":"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3"} Dec 01 20:02:56 crc kubenswrapper[4960]: I1201 20:02:56.283009 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"21d1f29c-ee4b-4009-8017-ee502c3cfab9","Type":"ContainerStarted","Data":"7026f6c36da5fc3be046075c10236ac5282cec17eabcea341248872260154650"} Dec 01 20:02:56 crc kubenswrapper[4960]: I1201 20:02:56.318325 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.318301032 podStartE2EDuration="2.318301032s" podCreationTimestamp="2025-12-01 20:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:02:56.299266683 +0000 UTC m=+1411.586758352" watchObservedRunningTime="2025-12-01 20:02:56.318301032 +0000 UTC m=+1411.605792711" Dec 01 20:02:56 crc kubenswrapper[4960]: I1201 20:02:56.564081 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.717945 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.720082 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.728507 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.728724 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zdz52" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.729255 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.774044 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.797853 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fdb5\" (UniqueName: \"kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.797956 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.798355 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.798468 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.899718 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.899803 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.899840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.899901 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fdb5\" (UniqueName: \"kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.900809 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.908274 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.908796 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:58 crc kubenswrapper[4960]: I1201 20:02:58.917870 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fdb5\" (UniqueName: \"kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5\") pod \"openstackclient\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.019377 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.020608 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.028031 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.117759 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.131939 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.140148 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 20:02:59 crc kubenswrapper[4960]: E1201 20:02:59.206000 4960 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 01 20:02:59 crc kubenswrapper[4960]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_ed36e5e7-37e1-4331-a899-9f25c95b63f2_0(30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38" Netns:"/var/run/netns/a1a99253-8278-4b00-8c83-8ac0015cd620" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38;K8S_POD_UID=ed36e5e7-37e1-4331-a899-9f25c95b63f2" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/ed36e5e7-37e1-4331-a899-9f25c95b63f2]: expected pod UID "ed36e5e7-37e1-4331-a899-9f25c95b63f2" but got "5f778c19-cf9c-4c12-811d-9bc33aae5e90" from Kube API Dec 01 20:02:59 crc kubenswrapper[4960]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 20:02:59 crc kubenswrapper[4960]: > Dec 01 20:02:59 crc kubenswrapper[4960]: E1201 20:02:59.206066 4960 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 01 20:02:59 crc kubenswrapper[4960]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_ed36e5e7-37e1-4331-a899-9f25c95b63f2_0(30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38" Netns:"/var/run/netns/a1a99253-8278-4b00-8c83-8ac0015cd620" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=30d1e993bf4aa6009da200193b825952141975621826cebde9d068e1a38c3c38;K8S_POD_UID=ed36e5e7-37e1-4331-a899-9f25c95b63f2" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/ed36e5e7-37e1-4331-a899-9f25c95b63f2]: expected pod UID "ed36e5e7-37e1-4331-a899-9f25c95b63f2" but got "5f778c19-cf9c-4c12-811d-9bc33aae5e90" from Kube API Dec 01 20:02:59 crc kubenswrapper[4960]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 01 20:02:59 crc kubenswrapper[4960]: > pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.213858 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.215745 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.215834 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pxzf\" (UniqueName: \"kubernetes.io/projected/5f778c19-cf9c-4c12-811d-9bc33aae5e90-kube-api-access-2pxzf\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.215924 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.315349 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.318912 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.319222 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.319268 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pxzf\" (UniqueName: \"kubernetes.io/projected/5f778c19-cf9c-4c12-811d-9bc33aae5e90-kube-api-access-2pxzf\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.319296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.319645 4960 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ed36e5e7-37e1-4331-a899-9f25c95b63f2" podUID="5f778c19-cf9c-4c12-811d-9bc33aae5e90" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.320604 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.324882 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-openstack-config-secret\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.326937 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f778c19-cf9c-4c12-811d-9bc33aae5e90-combined-ca-bundle\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.332692 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.344258 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pxzf\" (UniqueName: \"kubernetes.io/projected/5f778c19-cf9c-4c12-811d-9bc33aae5e90-kube-api-access-2pxzf\") pod \"openstackclient\" (UID: \"5f778c19-cf9c-4c12-811d-9bc33aae5e90\") " pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.421033 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config\") pod \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.421091 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret\") pod \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.421144 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fdb5\" (UniqueName: \"kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5\") pod \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.421345 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle\") pod \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\" (UID: \"ed36e5e7-37e1-4331-a899-9f25c95b63f2\") " Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.421669 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ed36e5e7-37e1-4331-a899-9f25c95b63f2" (UID: "ed36e5e7-37e1-4331-a899-9f25c95b63f2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.422059 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.425282 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed36e5e7-37e1-4331-a899-9f25c95b63f2" (UID: "ed36e5e7-37e1-4331-a899-9f25c95b63f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.425749 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ed36e5e7-37e1-4331-a899-9f25c95b63f2" (UID: "ed36e5e7-37e1-4331-a899-9f25c95b63f2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.425850 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5" (OuterVolumeSpecName: "kube-api-access-6fdb5") pod "ed36e5e7-37e1-4331-a899-9f25c95b63f2" (UID: "ed36e5e7-37e1-4331-a899-9f25c95b63f2"). InnerVolumeSpecName "kube-api-access-6fdb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.463724 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.524055 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.524095 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ed36e5e7-37e1-4331-a899-9f25c95b63f2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.524115 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fdb5\" (UniqueName: \"kubernetes.io/projected/ed36e5e7-37e1-4331-a899-9f25c95b63f2-kube-api-access-6fdb5\") on node \"crc\" DevicePath \"\"" Dec 01 20:02:59 crc kubenswrapper[4960]: W1201 20:02:59.952484 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f778c19_cf9c_4c12_811d_9bc33aae5e90.slice/crio-a7015f3a2b68d01e3ef02041015c7ca49e07a4fd0a01ca03144191a98ceb2418 WatchSource:0}: Error finding container a7015f3a2b68d01e3ef02041015c7ca49e07a4fd0a01ca03144191a98ceb2418: Status 404 returned error can't find the container with id a7015f3a2b68d01e3ef02041015c7ca49e07a4fd0a01ca03144191a98ceb2418 Dec 01 20:02:59 crc kubenswrapper[4960]: I1201 20:02:59.966373 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 01 20:03:00 crc kubenswrapper[4960]: I1201 20:03:00.327865 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5f778c19-cf9c-4c12-811d-9bc33aae5e90","Type":"ContainerStarted","Data":"a7015f3a2b68d01e3ef02041015c7ca49e07a4fd0a01ca03144191a98ceb2418"} Dec 01 20:03:00 crc kubenswrapper[4960]: I1201 20:03:00.327892 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 01 20:03:00 crc kubenswrapper[4960]: I1201 20:03:00.346611 4960 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ed36e5e7-37e1-4331-a899-9f25c95b63f2" podUID="5f778c19-cf9c-4c12-811d-9bc33aae5e90" Dec 01 20:03:01 crc kubenswrapper[4960]: I1201 20:03:01.343048 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed36e5e7-37e1-4331-a899-9f25c95b63f2" path="/var/lib/kubelet/pods/ed36e5e7-37e1-4331-a899-9f25c95b63f2/volumes" Dec 01 20:03:01 crc kubenswrapper[4960]: I1201 20:03:01.813604 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.138586 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.139038 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-central-agent" containerID="cri-o://14ceacb722cd22161637bdc4338bf3779f7a72183e4a239b6dc813e685c8a6cc" gracePeriod=30 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.139146 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="proxy-httpd" containerID="cri-o://f25fee3ed7a2dbc06c02b73be487d6baed181afa283327e83755925d4c1968ff" gracePeriod=30 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.139164 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-notification-agent" containerID="cri-o://689ffbbc20f65a6468ccba06177d0ed96525005202a475890c81476bf6b37765" gracePeriod=30 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.139397 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="sg-core" containerID="cri-o://d760d4e9de8c02a457eb77823cd01dc91c251788bfc958df8ea38e97db79399b" gracePeriod=30 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.151742 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.365841 4960 generic.go:334] "Generic (PLEG): container finished" podID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerID="f25fee3ed7a2dbc06c02b73be487d6baed181afa283327e83755925d4c1968ff" exitCode=0 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.366228 4960 generic.go:334] "Generic (PLEG): container finished" podID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerID="d760d4e9de8c02a457eb77823cd01dc91c251788bfc958df8ea38e97db79399b" exitCode=2 Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.365915 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerDied","Data":"f25fee3ed7a2dbc06c02b73be487d6baed181afa283327e83755925d4c1968ff"} Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.366274 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerDied","Data":"d760d4e9de8c02a457eb77823cd01dc91c251788bfc958df8ea38e97db79399b"} Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.530013 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-696549d5f9-6rmsc"] Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.535582 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.540865 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.540868 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.541991 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.543273 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696549d5f9-6rmsc"] Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.628942 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-combined-ca-bundle\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629187 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-log-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629245 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdrwq\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-kube-api-access-fdrwq\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629394 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-internal-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-run-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629536 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-public-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629751 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-etc-swift\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.629839 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-config-data\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732216 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-internal-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732270 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-run-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732298 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-public-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732330 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-etc-swift\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732355 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-config-data\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732406 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-combined-ca-bundle\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732472 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-log-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.732492 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdrwq\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-kube-api-access-fdrwq\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.733372 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-run-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.738752 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c99a2254-ef5a-47fd-8ed6-bc98e9671077-log-httpd\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.739395 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-config-data\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.739602 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-etc-swift\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.745897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-internal-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.746099 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-public-tls-certs\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.746275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c99a2254-ef5a-47fd-8ed6-bc98e9671077-combined-ca-bundle\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.756096 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdrwq\" (UniqueName: \"kubernetes.io/projected/c99a2254-ef5a-47fd-8ed6-bc98e9671077-kube-api-access-fdrwq\") pod \"swift-proxy-696549d5f9-6rmsc\" (UID: \"c99a2254-ef5a-47fd-8ed6-bc98e9671077\") " pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:03 crc kubenswrapper[4960]: I1201 20:03:03.924523 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.389568 4960 generic.go:334] "Generic (PLEG): container finished" podID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerID="689ffbbc20f65a6468ccba06177d0ed96525005202a475890c81476bf6b37765" exitCode=0 Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.389849 4960 generic.go:334] "Generic (PLEG): container finished" podID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerID="14ceacb722cd22161637bdc4338bf3779f7a72183e4a239b6dc813e685c8a6cc" exitCode=0 Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.389670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerDied","Data":"689ffbbc20f65a6468ccba06177d0ed96525005202a475890c81476bf6b37765"} Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.389893 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerDied","Data":"14ceacb722cd22161637bdc4338bf3779f7a72183e4a239b6dc813e685c8a6cc"} Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.496098 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-696549d5f9-6rmsc"] Dec 01 20:03:04 crc kubenswrapper[4960]: W1201 20:03:04.500675 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc99a2254_ef5a_47fd_8ed6_bc98e9671077.slice/crio-a319e940e2f155b133c028ce9d081f1d8756e2df9356ea8cd6d74022a2196900 WatchSource:0}: Error finding container a319e940e2f155b133c028ce9d081f1d8756e2df9356ea8cd6d74022a2196900: Status 404 returned error can't find the container with id a319e940e2f155b133c028ce9d081f1d8756e2df9356ea8cd6d74022a2196900 Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.685986 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759004 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759081 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759149 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759286 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759332 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759356 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759424 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkx5j\" (UniqueName: \"kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j\") pod \"9117e004-0aac-4807-9c01-5c35cdc1d72a\" (UID: \"9117e004-0aac-4807-9c01-5c35cdc1d72a\") " Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.759904 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.760006 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.765495 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts" (OuterVolumeSpecName: "scripts") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.801095 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j" (OuterVolumeSpecName: "kube-api-access-kkx5j") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "kube-api-access-kkx5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.809055 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.861587 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.861624 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.861635 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.861647 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkx5j\" (UniqueName: \"kubernetes.io/projected/9117e004-0aac-4807-9c01-5c35cdc1d72a-kube-api-access-kkx5j\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.861661 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9117e004-0aac-4807-9c01-5c35cdc1d72a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.870064 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.896732 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data" (OuterVolumeSpecName: "config-data") pod "9117e004-0aac-4807-9c01-5c35cdc1d72a" (UID: "9117e004-0aac-4807-9c01-5c35cdc1d72a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.963174 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:04 crc kubenswrapper[4960]: I1201 20:03:04.963209 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9117e004-0aac-4807-9c01-5c35cdc1d72a-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.416754 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9117e004-0aac-4807-9c01-5c35cdc1d72a","Type":"ContainerDied","Data":"79431ae1511cc5a17e93fb95f001b2c60d759de6a2718dd3322abbf754a0eaae"} Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.417040 4960 scope.go:117] "RemoveContainer" containerID="f25fee3ed7a2dbc06c02b73be487d6baed181afa283327e83755925d4c1968ff" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.417208 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.433466 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696549d5f9-6rmsc" event={"ID":"c99a2254-ef5a-47fd-8ed6-bc98e9671077","Type":"ContainerStarted","Data":"bd2ba31736bbd359ab4aeed1f8bfaa322a16e0bd0ce7f4fabb353a053bb9b329"} Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.433530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696549d5f9-6rmsc" event={"ID":"c99a2254-ef5a-47fd-8ed6-bc98e9671077","Type":"ContainerStarted","Data":"b8923cac64f0f64ba580464d26dbb449e723a81164bb4c2eea2c5e21547b0827"} Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.433543 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-696549d5f9-6rmsc" event={"ID":"c99a2254-ef5a-47fd-8ed6-bc98e9671077","Type":"ContainerStarted","Data":"a319e940e2f155b133c028ce9d081f1d8756e2df9356ea8cd6d74022a2196900"} Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.434685 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.434711 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.444524 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.454951 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.460778 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-696549d5f9-6rmsc" podStartSLOduration=2.460752088 podStartE2EDuration="2.460752088s" podCreationTimestamp="2025-12-01 20:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:05.452506394 +0000 UTC m=+1420.739998093" watchObservedRunningTime="2025-12-01 20:03:05.460752088 +0000 UTC m=+1420.748243757" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.511627 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:05 crc kubenswrapper[4960]: E1201 20:03:05.512253 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-central-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512270 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-central-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: E1201 20:03:05.512284 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-notification-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512290 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-notification-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: E1201 20:03:05.512310 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="proxy-httpd" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512317 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="proxy-httpd" Dec 01 20:03:05 crc kubenswrapper[4960]: E1201 20:03:05.512336 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="sg-core" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512341 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="sg-core" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512595 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-notification-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512610 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="sg-core" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512623 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="ceilometer-central-agent" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.512637 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" containerName="proxy-httpd" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.515092 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.517597 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.517847 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.524427 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577479 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577578 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577631 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577679 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gv2f\" (UniqueName: \"kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577716 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.577742 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679707 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679838 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gv2f\" (UniqueName: \"kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679858 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679882 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.679963 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.682242 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.682348 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.688430 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.689175 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.691926 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.702629 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.707134 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gv2f\" (UniqueName: \"kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f\") pod \"ceilometer-0\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " pod="openstack/ceilometer-0" Dec 01 20:03:05 crc kubenswrapper[4960]: I1201 20:03:05.836298 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:07 crc kubenswrapper[4960]: I1201 20:03:07.336825 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9117e004-0aac-4807-9c01-5c35cdc1d72a" path="/var/lib/kubelet/pods/9117e004-0aac-4807-9c01-5c35cdc1d72a/volumes" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.180684 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8sh5f"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.182642 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.197013 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8sh5f"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.276290 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8sg2\" (UniqueName: \"kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.276362 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.304628 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jb2lh"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.306274 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.349500 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jb2lh"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.380623 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.381180 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gtgg\" (UniqueName: \"kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.381282 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8sg2\" (UniqueName: \"kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.381385 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.383648 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.400082 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0a5a-account-create-update-j28nl"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.401700 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.409817 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.414692 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8sg2\" (UniqueName: \"kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2\") pod \"nova-api-db-create-8sh5f\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.439107 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-wrksp"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.440743 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.452169 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wrksp"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.465218 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0a5a-account-create-update-j28nl"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.483598 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.483765 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.483934 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gtgg\" (UniqueName: \"kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.484019 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tpxp\" (UniqueName: \"kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.484346 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.503435 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gtgg\" (UniqueName: \"kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg\") pod \"nova-cell0-db-create-jb2lh\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.514258 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.585855 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tpxp\" (UniqueName: \"kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.585963 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.586004 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.586028 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb579\" (UniqueName: \"kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.586177 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-741e-account-create-update-pw8w9"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.586932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.588078 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.591425 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.608126 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tpxp\" (UniqueName: \"kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp\") pod \"nova-api-0a5a-account-create-update-j28nl\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.610986 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-741e-account-create-update-pw8w9"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.632468 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.689214 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.689801 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.689946 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.690072 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb579\" (UniqueName: \"kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.691034 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.710969 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb579\" (UniqueName: \"kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579\") pod \"nova-cell1-db-create-wrksp\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.784172 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.793702 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.793842 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.794534 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.796182 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.796963 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ab80-account-create-update-bvzrn"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.798966 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.803177 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.821834 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr\") pod \"nova-cell0-741e-account-create-update-pw8w9\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.848486 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ab80-account-create-update-bvzrn"] Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.895563 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvr9k\" (UniqueName: \"kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.895757 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.960283 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.998027 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.998177 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvr9k\" (UniqueName: \"kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:09 crc kubenswrapper[4960]: I1201 20:03:09.998765 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.017654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvr9k\" (UniqueName: \"kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k\") pod \"nova-cell1-ab80-account-create-update-bvzrn\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.185484 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.286440 4960 scope.go:117] "RemoveContainer" containerID="d760d4e9de8c02a457eb77823cd01dc91c251788bfc958df8ea38e97db79399b" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.523087 4960 generic.go:334] "Generic (PLEG): container finished" podID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerID="56a9a27a582bb6de3091a6c2ec883fc77b2ea4f689f5ace6a4ccc4b8f8df2c95" exitCode=137 Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.523709 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerDied","Data":"56a9a27a582bb6de3091a6c2ec883fc77b2ea4f689f5ace6a4ccc4b8f8df2c95"} Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.546704 4960 scope.go:117] "RemoveContainer" containerID="689ffbbc20f65a6468ccba06177d0ed96525005202a475890c81476bf6b37765" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.800209 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.842150 4960 scope.go:117] "RemoveContainer" containerID="14ceacb722cd22161637bdc4338bf3779f7a72183e4a239b6dc813e685c8a6cc" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.919987 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920086 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920151 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920196 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920278 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920304 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.920373 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz726\" (UniqueName: \"kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726\") pod \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\" (UID: \"5fc666b0-f236-45ee-95ec-0c35cd3df71b\") " Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.921644 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.922062 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs" (OuterVolumeSpecName: "logs") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.926082 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726" (OuterVolumeSpecName: "kube-api-access-pz726") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "kube-api-access-pz726". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.936330 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:10 crc kubenswrapper[4960]: I1201 20:03:10.939817 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts" (OuterVolumeSpecName: "scripts") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.006332 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029167 4960 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5fc666b0-f236-45ee-95ec-0c35cd3df71b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029193 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029202 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz726\" (UniqueName: \"kubernetes.io/projected/5fc666b0-f236-45ee-95ec-0c35cd3df71b-kube-api-access-pz726\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029211 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029219 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc666b0-f236-45ee-95ec-0c35cd3df71b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.029226 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.071103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data" (OuterVolumeSpecName: "config-data") pod "5fc666b0-f236-45ee-95ec-0c35cd3df71b" (UID: "5fc666b0-f236-45ee-95ec-0c35cd3df71b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.131074 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc666b0-f236-45ee-95ec-0c35cd3df71b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.159310 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ab80-account-create-update-bvzrn"] Dec 01 20:03:11 crc kubenswrapper[4960]: W1201 20:03:11.400371 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd0349dc_1091_4e8f_921f_71a349b674d2.slice/crio-6a4bb08a63a51053ddfd7a634f043805750555a07c6547224602c5380bbbb0c4 WatchSource:0}: Error finding container 6a4bb08a63a51053ddfd7a634f043805750555a07c6547224602c5380bbbb0c4: Status 404 returned error can't find the container with id 6a4bb08a63a51053ddfd7a634f043805750555a07c6547224602c5380bbbb0c4 Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.404234 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.415989 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0a5a-account-create-update-j28nl"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.554868 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-wrksp"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.570181 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wrksp" event={"ID":"3e263113-788c-4b96-bb25-5bc58b79a29a","Type":"ContainerStarted","Data":"cb2bb95a9a8e9595893c1e68ad439b039500a2d86d42c721baea251f9f1629e5"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.576100 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5fc666b0-f236-45ee-95ec-0c35cd3df71b","Type":"ContainerDied","Data":"e2ac69a6cf031a148b3a22d6838c0a7b01f20c9399c5974896c72414425f35c9"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.576238 4960 scope.go:117] "RemoveContainer" containerID="56a9a27a582bb6de3091a6c2ec883fc77b2ea4f689f5ace6a4ccc4b8f8df2c95" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.576352 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.585081 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerStarted","Data":"6a4bb08a63a51053ddfd7a634f043805750555a07c6547224602c5380bbbb0c4"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.593519 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"5f778c19-cf9c-4c12-811d-9bc33aae5e90","Type":"ContainerStarted","Data":"281a86c59baa55c69dd6544b0fb80eef16cba31546b22555e35298639c3d7de0"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.604228 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0a5a-account-create-update-j28nl" event={"ID":"98bcf1ff-8858-41e7-8c27-e0e6229b853d","Type":"ContainerStarted","Data":"3718425a32a3ca0fd708e70990024b3d3b443eda61d42dc1c1af2b9ee4acbd1d"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.617329 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" event={"ID":"ac682301-3973-4d8d-9091-ab0a3e2ba851","Type":"ContainerStarted","Data":"d6d87e3675c17909fba324aaad98765ff94ec8c0a15823449f65c7edd7f4e5ba"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.617388 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" event={"ID":"ac682301-3973-4d8d-9091-ab0a3e2ba851","Type":"ContainerStarted","Data":"ce1fa7ab69e2fec5638ce0927703a022d6d02e86412baf9178f9395e0a542e30"} Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.621112 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.9010435540000001 podStartE2EDuration="12.621083627s" podCreationTimestamp="2025-12-01 20:02:59 +0000 UTC" firstStartedPulling="2025-12-01 20:02:59.955165357 +0000 UTC m=+1415.242657026" lastFinishedPulling="2025-12-01 20:03:10.67520543 +0000 UTC m=+1425.962697099" observedRunningTime="2025-12-01 20:03:11.610561758 +0000 UTC m=+1426.898053427" watchObservedRunningTime="2025-12-01 20:03:11.621083627 +0000 UTC m=+1426.908575296" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.654467 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" podStartSLOduration=2.65444493 podStartE2EDuration="2.65444493s" podCreationTimestamp="2025-12-01 20:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:11.632879816 +0000 UTC m=+1426.920371475" watchObservedRunningTime="2025-12-01 20:03:11.65444493 +0000 UTC m=+1426.941936599" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.666332 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-741e-account-create-update-pw8w9"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.674854 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jb2lh"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.682754 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8sh5f"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.765431 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.780260 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.782370 4960 scope.go:117] "RemoveContainer" containerID="5fdbe480a5d2707c8943a01ec1e91a1db93049ba8e23331ebfb06202ec12c02f" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.793335 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:03:11 crc kubenswrapper[4960]: E1201 20:03:11.793850 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.793868 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api" Dec 01 20:03:11 crc kubenswrapper[4960]: E1201 20:03:11.793890 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api-log" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.793897 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api-log" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.794074 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api-log" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.794090 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" containerName="cinder-api" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.795332 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.800161 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.800349 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.800651 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.803580 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846138 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2phk8\" (UniqueName: \"kubernetes.io/projected/b45703c2-5b71-47a4-835a-cee374c228b7-kube-api-access-2phk8\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846185 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846207 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b45703c2-5b71-47a4-835a-cee374c228b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846298 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846719 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846771 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-scripts\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846811 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45703c2-5b71-47a4-835a-cee374c228b7-logs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.846852 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948533 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2phk8\" (UniqueName: \"kubernetes.io/projected/b45703c2-5b71-47a4-835a-cee374c228b7-kube-api-access-2phk8\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948582 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948610 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948631 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b45703c2-5b71-47a4-835a-cee374c228b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948721 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948778 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-scripts\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948812 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45703c2-5b71-47a4-835a-cee374c228b7-logs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.948851 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.949337 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b45703c2-5b71-47a4-835a-cee374c228b7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.949652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45703c2-5b71-47a4-835a-cee374c228b7-logs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.955415 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data-custom\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.956834 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.956945 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-config-data\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.957314 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.957707 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.957859 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b45703c2-5b71-47a4-835a-cee374c228b7-scripts\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:11 crc kubenswrapper[4960]: I1201 20:03:11.976960 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2phk8\" (UniqueName: \"kubernetes.io/projected/b45703c2-5b71-47a4-835a-cee374c228b7-kube-api-access-2phk8\") pod \"cinder-api-0\" (UID: \"b45703c2-5b71-47a4-835a-cee374c228b7\") " pod="openstack/cinder-api-0" Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.128072 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.592927 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 01 20:03:12 crc kubenswrapper[4960]: W1201 20:03:12.596011 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb45703c2_5b71_47a4_835a_cee374c228b7.slice/crio-f8700b77e6231ea0d7b2faac5b5faac936560812ccf7f6a3723eda1d5a7eac9f WatchSource:0}: Error finding container f8700b77e6231ea0d7b2faac5b5faac936560812ccf7f6a3723eda1d5a7eac9f: Status 404 returned error can't find the container with id f8700b77e6231ea0d7b2faac5b5faac936560812ccf7f6a3723eda1d5a7eac9f Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.629636 4960 generic.go:334] "Generic (PLEG): container finished" podID="2f044d4c-9a53-4a0b-b8a8-f27eb728414b" containerID="8be31e18d2bc3a186cafd70c832374703a971d9d22da5cde6ca11ce540f4d38e" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.629711 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" event={"ID":"2f044d4c-9a53-4a0b-b8a8-f27eb728414b","Type":"ContainerDied","Data":"8be31e18d2bc3a186cafd70c832374703a971d9d22da5cde6ca11ce540f4d38e"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.629745 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" event={"ID":"2f044d4c-9a53-4a0b-b8a8-f27eb728414b","Type":"ContainerStarted","Data":"dcd311a546306676a39b6cf6a637833bfefdbce2640f4189cc32b37a5fc331a4"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.632466 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c257117-da7a-4840-944e-ff57c68838d6" containerID="f0b0f2e0596daff6ec7c5ad7f83a5b8d7c159d26338123d3e8ab1361d21f9c4d" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.632581 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jb2lh" event={"ID":"4c257117-da7a-4840-944e-ff57c68838d6","Type":"ContainerDied","Data":"f0b0f2e0596daff6ec7c5ad7f83a5b8d7c159d26338123d3e8ab1361d21f9c4d"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.632600 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jb2lh" event={"ID":"4c257117-da7a-4840-944e-ff57c68838d6","Type":"ContainerStarted","Data":"c6a7a83278c126335aa100e324ce7a6da34436db6eaf1bc0d3cee88a086f3034"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.643053 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerStarted","Data":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.650344 4960 generic.go:334] "Generic (PLEG): container finished" podID="3e263113-788c-4b96-bb25-5bc58b79a29a" containerID="42d4361844a02929c43bfd8fbca28a5c7baba6d859ae982a6ce92b0d44070742" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.650448 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wrksp" event={"ID":"3e263113-788c-4b96-bb25-5bc58b79a29a","Type":"ContainerDied","Data":"42d4361844a02929c43bfd8fbca28a5c7baba6d859ae982a6ce92b0d44070742"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.653456 4960 generic.go:334] "Generic (PLEG): container finished" podID="494b9d4e-5308-4954-92f3-39e82ff83380" containerID="76c1e247963966f85a2bbde937758f7c67f0ed6a5dbc5ab1cd889fff29d08b07" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.653568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8sh5f" event={"ID":"494b9d4e-5308-4954-92f3-39e82ff83380","Type":"ContainerDied","Data":"76c1e247963966f85a2bbde937758f7c67f0ed6a5dbc5ab1cd889fff29d08b07"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.653596 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8sh5f" event={"ID":"494b9d4e-5308-4954-92f3-39e82ff83380","Type":"ContainerStarted","Data":"f869735fc64319e5be5818018175b619156e9c713f08d555bf455974d1fd9b82"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.656329 4960 generic.go:334] "Generic (PLEG): container finished" podID="ac682301-3973-4d8d-9091-ab0a3e2ba851" containerID="d6d87e3675c17909fba324aaad98765ff94ec8c0a15823449f65c7edd7f4e5ba" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.656419 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" event={"ID":"ac682301-3973-4d8d-9091-ab0a3e2ba851","Type":"ContainerDied","Data":"d6d87e3675c17909fba324aaad98765ff94ec8c0a15823449f65c7edd7f4e5ba"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.665242 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b45703c2-5b71-47a4-835a-cee374c228b7","Type":"ContainerStarted","Data":"f8700b77e6231ea0d7b2faac5b5faac936560812ccf7f6a3723eda1d5a7eac9f"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.669896 4960 generic.go:334] "Generic (PLEG): container finished" podID="98bcf1ff-8858-41e7-8c27-e0e6229b853d" containerID="d114c4107a031ac1b3d9ac0420107c3bd595c0d075c5c6ba58438a416c43acc6" exitCode=0 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.670215 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0a5a-account-create-update-j28nl" event={"ID":"98bcf1ff-8858-41e7-8c27-e0e6229b853d","Type":"ContainerDied","Data":"d114c4107a031ac1b3d9ac0420107c3bd595c0d075c5c6ba58438a416c43acc6"} Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.989884 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.990464 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-log" containerID="cri-o://f9b86a25e5c116940aa360b7d179569e5c387f9371bca29da93408cef36aa980" gracePeriod=30 Dec 01 20:03:12 crc kubenswrapper[4960]: I1201 20:03:12.990992 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-httpd" containerID="cri-o://1d8f606812105aa624867384d53a722d0010b121426d44171d1d0b7694081716" gracePeriod=30 Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.358026 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc666b0-f236-45ee-95ec-0c35cd3df71b" path="/var/lib/kubelet/pods/5fc666b0-f236-45ee-95ec-0c35cd3df71b/volumes" Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.481676 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.681811 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b45703c2-5b71-47a4-835a-cee374c228b7","Type":"ContainerStarted","Data":"bdcdd40291dbaf1e30b28b1487994ac2dbf160e479e1f4bd49852344e64052bb"} Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.687436 4960 generic.go:334] "Generic (PLEG): container finished" podID="3ed64227-884f-41f4-8534-1bc160d45726" containerID="f9b86a25e5c116940aa360b7d179569e5c387f9371bca29da93408cef36aa980" exitCode=143 Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.687502 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerDied","Data":"f9b86a25e5c116940aa360b7d179569e5c387f9371bca29da93408cef36aa980"} Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.693713 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerStarted","Data":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.871503 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.871577 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.936074 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:13 crc kubenswrapper[4960]: I1201 20:03:13.939848 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-696549d5f9-6rmsc" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.248964 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.310028 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts\") pod \"494b9d4e-5308-4954-92f3-39e82ff83380\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.310716 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8sg2\" (UniqueName: \"kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2\") pod \"494b9d4e-5308-4954-92f3-39e82ff83380\" (UID: \"494b9d4e-5308-4954-92f3-39e82ff83380\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.314248 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "494b9d4e-5308-4954-92f3-39e82ff83380" (UID: "494b9d4e-5308-4954-92f3-39e82ff83380"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.321391 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2" (OuterVolumeSpecName: "kube-api-access-p8sg2") pod "494b9d4e-5308-4954-92f3-39e82ff83380" (UID: "494b9d4e-5308-4954-92f3-39e82ff83380"). InnerVolumeSpecName "kube-api-access-p8sg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.413495 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/494b9d4e-5308-4954-92f3-39e82ff83380-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.413543 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8sg2\" (UniqueName: \"kubernetes.io/projected/494b9d4e-5308-4954-92f3-39e82ff83380-kube-api-access-p8sg2\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.627334 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.628592 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.654941 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.661834 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.678293 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.710927 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-wrksp" event={"ID":"3e263113-788c-4b96-bb25-5bc58b79a29a","Type":"ContainerDied","Data":"cb2bb95a9a8e9595893c1e68ad439b039500a2d86d42c721baea251f9f1629e5"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.712356 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb2bb95a9a8e9595893c1e68ad439b039500a2d86d42c721baea251f9f1629e5" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.711035 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-wrksp" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.716107 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8sh5f" event={"ID":"494b9d4e-5308-4954-92f3-39e82ff83380","Type":"ContainerDied","Data":"f869735fc64319e5be5818018175b619156e9c713f08d555bf455974d1fd9b82"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.719611 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f869735fc64319e5be5818018175b619156e9c713f08d555bf455974d1fd9b82" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.716385 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8sh5f" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726164 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tpxp\" (UniqueName: \"kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp\") pod \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726239 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts\") pod \"3e263113-788c-4b96-bb25-5bc58b79a29a\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726264 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb579\" (UniqueName: \"kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579\") pod \"3e263113-788c-4b96-bb25-5bc58b79a29a\" (UID: \"3e263113-788c-4b96-bb25-5bc58b79a29a\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726346 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvr9k\" (UniqueName: \"kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k\") pod \"ac682301-3973-4d8d-9091-ab0a3e2ba851\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726409 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts\") pod \"ac682301-3973-4d8d-9091-ab0a3e2ba851\" (UID: \"ac682301-3973-4d8d-9091-ab0a3e2ba851\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726444 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts\") pod \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\" (UID: \"98bcf1ff-8858-41e7-8c27-e0e6229b853d\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726482 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr\") pod \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726536 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts\") pod \"4c257117-da7a-4840-944e-ff57c68838d6\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726557 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts\") pod \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\" (UID: \"2f044d4c-9a53-4a0b-b8a8-f27eb728414b\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.726612 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gtgg\" (UniqueName: \"kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg\") pod \"4c257117-da7a-4840-944e-ff57c68838d6\" (UID: \"4c257117-da7a-4840-944e-ff57c68838d6\") " Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.727210 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac682301-3973-4d8d-9091-ab0a3e2ba851" (UID: "ac682301-3973-4d8d-9091-ab0a3e2ba851"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.733813 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f044d4c-9a53-4a0b-b8a8-f27eb728414b" (UID: "2f044d4c-9a53-4a0b-b8a8-f27eb728414b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.734629 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e263113-788c-4b96-bb25-5bc58b79a29a" (UID: "3e263113-788c-4b96-bb25-5bc58b79a29a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.735456 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b45703c2-5b71-47a4-835a-cee374c228b7","Type":"ContainerStarted","Data":"d568ac533e29ed6d9f35f099b64d9adca5c7de96f7655192c29dd1a21a00bcf4"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.736027 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr" (OuterVolumeSpecName: "kube-api-access-6rxgr") pod "2f044d4c-9a53-4a0b-b8a8-f27eb728414b" (UID: "2f044d4c-9a53-4a0b-b8a8-f27eb728414b"). InnerVolumeSpecName "kube-api-access-6rxgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.737387 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98bcf1ff-8858-41e7-8c27-e0e6229b853d" (UID: "98bcf1ff-8858-41e7-8c27-e0e6229b853d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.737696 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c257117-da7a-4840-944e-ff57c68838d6" (UID: "4c257117-da7a-4840-944e-ff57c68838d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.738050 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.741020 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg" (OuterVolumeSpecName: "kube-api-access-7gtgg") pod "4c257117-da7a-4840-944e-ff57c68838d6" (UID: "4c257117-da7a-4840-944e-ff57c68838d6"). InnerVolumeSpecName "kube-api-access-7gtgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.741247 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k" (OuterVolumeSpecName: "kube-api-access-mvr9k") pod "ac682301-3973-4d8d-9091-ab0a3e2ba851" (UID: "ac682301-3973-4d8d-9091-ab0a3e2ba851"). InnerVolumeSpecName "kube-api-access-mvr9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.746472 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" event={"ID":"ac682301-3973-4d8d-9091-ab0a3e2ba851","Type":"ContainerDied","Data":"ce1fa7ab69e2fec5638ce0927703a022d6d02e86412baf9178f9395e0a542e30"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.746518 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce1fa7ab69e2fec5638ce0927703a022d6d02e86412baf9178f9395e0a542e30" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.746489 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ab80-account-create-update-bvzrn" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.749029 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp" (OuterVolumeSpecName: "kube-api-access-5tpxp") pod "98bcf1ff-8858-41e7-8c27-e0e6229b853d" (UID: "98bcf1ff-8858-41e7-8c27-e0e6229b853d"). InnerVolumeSpecName "kube-api-access-5tpxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.749341 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0a5a-account-create-update-j28nl" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.749444 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0a5a-account-create-update-j28nl" event={"ID":"98bcf1ff-8858-41e7-8c27-e0e6229b853d","Type":"ContainerDied","Data":"3718425a32a3ca0fd708e70990024b3d3b443eda61d42dc1c1af2b9ee4acbd1d"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.749489 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3718425a32a3ca0fd708e70990024b3d3b443eda61d42dc1c1af2b9ee4acbd1d" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.752230 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.752273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-741e-account-create-update-pw8w9" event={"ID":"2f044d4c-9a53-4a0b-b8a8-f27eb728414b","Type":"ContainerDied","Data":"dcd311a546306676a39b6cf6a637833bfefdbce2640f4189cc32b37a5fc331a4"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.752326 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcd311a546306676a39b6cf6a637833bfefdbce2640f4189cc32b37a5fc331a4" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.755531 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jb2lh" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.755575 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jb2lh" event={"ID":"4c257117-da7a-4840-944e-ff57c68838d6","Type":"ContainerDied","Data":"c6a7a83278c126335aa100e324ce7a6da34436db6eaf1bc0d3cee88a086f3034"} Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.755606 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6a7a83278c126335aa100e324ce7a6da34436db6eaf1bc0d3cee88a086f3034" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.761578 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579" (OuterVolumeSpecName: "kube-api-access-xb579") pod "3e263113-788c-4b96-bb25-5bc58b79a29a" (UID: "3e263113-788c-4b96-bb25-5bc58b79a29a"). InnerVolumeSpecName "kube-api-access-xb579". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.810586 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.810562988 podStartE2EDuration="3.810562988s" podCreationTimestamp="2025-12-01 20:03:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:14.770286867 +0000 UTC m=+1430.057778536" watchObservedRunningTime="2025-12-01 20:03:14.810562988 +0000 UTC m=+1430.098054657" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828643 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98bcf1ff-8858-41e7-8c27-e0e6229b853d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828666 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rxgr\" (UniqueName: \"kubernetes.io/projected/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-kube-api-access-6rxgr\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828676 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c257117-da7a-4840-944e-ff57c68838d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828686 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f044d4c-9a53-4a0b-b8a8-f27eb728414b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828696 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gtgg\" (UniqueName: \"kubernetes.io/projected/4c257117-da7a-4840-944e-ff57c68838d6-kube-api-access-7gtgg\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828704 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tpxp\" (UniqueName: \"kubernetes.io/projected/98bcf1ff-8858-41e7-8c27-e0e6229b853d-kube-api-access-5tpxp\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828712 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e263113-788c-4b96-bb25-5bc58b79a29a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828722 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb579\" (UniqueName: \"kubernetes.io/projected/3e263113-788c-4b96-bb25-5bc58b79a29a-kube-api-access-xb579\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828731 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvr9k\" (UniqueName: \"kubernetes.io/projected/ac682301-3973-4d8d-9091-ab0a3e2ba851-kube-api-access-mvr9k\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:14 crc kubenswrapper[4960]: I1201 20:03:14.828739 4960 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac682301-3973-4d8d-9091-ab0a3e2ba851-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:15 crc kubenswrapper[4960]: I1201 20:03:15.768568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerStarted","Data":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.274682 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.274916 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-log" containerID="cri-o://8fd8e20abe45a1b5aa11c7209951adfdad2e47d20e9cb8f61bb109149a87f0b7" gracePeriod=30 Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.275873 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-httpd" containerID="cri-o://def26066f8c718e9c7360c95a45942c4818705caa17fa1ccc96fee88aec073f7" gracePeriod=30 Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.778798 4960 generic.go:334] "Generic (PLEG): container finished" podID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerID="8fd8e20abe45a1b5aa11c7209951adfdad2e47d20e9cb8f61bb109149a87f0b7" exitCode=143 Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.778857 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerDied","Data":"8fd8e20abe45a1b5aa11c7209951adfdad2e47d20e9cb8f61bb109149a87f0b7"} Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.781781 4960 generic.go:334] "Generic (PLEG): container finished" podID="3ed64227-884f-41f4-8534-1bc160d45726" containerID="1d8f606812105aa624867384d53a722d0010b121426d44171d1d0b7694081716" exitCode=0 Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.781817 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerDied","Data":"1d8f606812105aa624867384d53a722d0010b121426d44171d1d0b7694081716"} Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.781847 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ed64227-884f-41f4-8534-1bc160d45726","Type":"ContainerDied","Data":"7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e"} Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.781858 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f25f74a38d4a89f7261035d4c3dbba741102896570e093ad0ae2c1a08490d4e" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.781867 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.873788 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.873837 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.873866 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874072 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874101 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bpgx\" (UniqueName: \"kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874155 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874264 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874684 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874780 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs" (OuterVolumeSpecName: "logs") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.874942 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle\") pod \"3ed64227-884f-41f4-8534-1bc160d45726\" (UID: \"3ed64227-884f-41f4-8534-1bc160d45726\") " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.883934 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.883976 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ed64227-884f-41f4-8534-1bc160d45726-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.885442 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts" (OuterVolumeSpecName: "scripts") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.911428 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx" (OuterVolumeSpecName: "kube-api-access-9bpgx") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "kube-api-access-9bpgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.929374 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52" (OuterVolumeSpecName: "glance") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.986723 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.986960 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") on node \"crc\" " Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.986973 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bpgx\" (UniqueName: \"kubernetes.io/projected/3ed64227-884f-41f4-8534-1bc160d45726-kube-api-access-9bpgx\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:16 crc kubenswrapper[4960]: I1201 20:03:16.987511 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.004470 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.016225 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.016490 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52") on node "crc" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.019967 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data" (OuterVolumeSpecName: "config-data") pod "3ed64227-884f-41f4-8534-1bc160d45726" (UID: "3ed64227-884f-41f4-8534-1bc160d45726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.088928 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.088968 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.088978 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.088989 4960 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ed64227-884f-41f4-8534-1bc160d45726-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792264 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792646 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="proxy-httpd" containerID="cri-o://a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" gracePeriod=30 Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792735 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="sg-core" containerID="cri-o://93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" gracePeriod=30 Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792779 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-notification-agent" containerID="cri-o://e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" gracePeriod=30 Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792360 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-central-agent" containerID="cri-o://14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" gracePeriod=30 Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792283 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerStarted","Data":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.792988 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.833523 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.930744677 podStartE2EDuration="12.833500307s" podCreationTimestamp="2025-12-01 20:03:05 +0000 UTC" firstStartedPulling="2025-12-01 20:03:11.418268651 +0000 UTC m=+1426.705760320" lastFinishedPulling="2025-12-01 20:03:17.321024271 +0000 UTC m=+1432.608515950" observedRunningTime="2025-12-01 20:03:17.819052024 +0000 UTC m=+1433.106543693" watchObservedRunningTime="2025-12-01 20:03:17.833500307 +0000 UTC m=+1433.120991976" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.862904 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.884817 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.899689 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.900575 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="494b9d4e-5308-4954-92f3-39e82ff83380" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.900599 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="494b9d4e-5308-4954-92f3-39e82ff83380" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.900613 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e263113-788c-4b96-bb25-5bc58b79a29a" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.900621 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e263113-788c-4b96-bb25-5bc58b79a29a" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.900664 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c257117-da7a-4840-944e-ff57c68838d6" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.900673 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c257117-da7a-4840-944e-ff57c68838d6" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.900688 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-httpd" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.900696 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-httpd" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.900761 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac682301-3973-4d8d-9091-ab0a3e2ba851" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.900771 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac682301-3973-4d8d-9091-ab0a3e2ba851" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.901195 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98bcf1ff-8858-41e7-8c27-e0e6229b853d" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901242 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="98bcf1ff-8858-41e7-8c27-e0e6229b853d" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.901323 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f044d4c-9a53-4a0b-b8a8-f27eb728414b" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901337 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f044d4c-9a53-4a0b-b8a8-f27eb728414b" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: E1201 20:03:17.901356 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-log" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901364 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-log" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901740 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c257117-da7a-4840-944e-ff57c68838d6" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901766 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-httpd" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901781 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="98bcf1ff-8858-41e7-8c27-e0e6229b853d" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901813 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="494b9d4e-5308-4954-92f3-39e82ff83380" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901828 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac682301-3973-4d8d-9091-ab0a3e2ba851" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901845 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e263113-788c-4b96-bb25-5bc58b79a29a" containerName="mariadb-database-create" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901860 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed64227-884f-41f4-8534-1bc160d45726" containerName="glance-log" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.901890 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f044d4c-9a53-4a0b-b8a8-f27eb728414b" containerName="mariadb-account-create-update" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.904633 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.910545 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.910740 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 01 20:03:17 crc kubenswrapper[4960]: I1201 20:03:17.925192 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007766 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007819 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007845 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007866 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-logs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007914 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkwnh\" (UniqueName: \"kubernetes.io/projected/5a341d11-27da-4def-a478-c878a69e2e6b-kube-api-access-kkwnh\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.007933 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.008004 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.008036 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109428 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkwnh\" (UniqueName: \"kubernetes.io/projected/5a341d11-27da-4def-a478-c878a69e2e6b-kube-api-access-kkwnh\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109475 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109558 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109593 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109673 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109697 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.109717 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-logs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.112259 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.113452 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a341d11-27da-4def-a478-c878a69e2e6b-logs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.117446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.117541 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.117759 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.117847 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a341d11-27da-4def-a478-c878a69e2e6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.117973 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.118005 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/91baf06d123dd91aed63649f36b06b6d82695580db5bb21a6dba903e67f7a922/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.127299 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkwnh\" (UniqueName: \"kubernetes.io/projected/5a341d11-27da-4def-a478-c878a69e2e6b-kube-api-access-kkwnh\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.160465 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b676a8ee-b56e-4be6-9a0a-6f5f9f57bf52\") pod \"glance-default-external-api-0\" (UID: \"5a341d11-27da-4def-a478-c878a69e2e6b\") " pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.321103 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.777246 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823089 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823162 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823192 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823232 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823250 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823289 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823314 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gv2f\" (UniqueName: \"kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f\") pod \"cd0349dc-1091-4e8f-921f-71a349b674d2\" (UID: \"cd0349dc-1091-4e8f-921f-71a349b674d2\") " Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.823611 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.824011 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.824434 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834486 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f" (OuterVolumeSpecName: "kube-api-access-4gv2f") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "kube-api-access-4gv2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834548 4960 generic.go:334] "Generic (PLEG): container finished" podID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" exitCode=0 Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834578 4960 generic.go:334] "Generic (PLEG): container finished" podID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" exitCode=2 Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834601 4960 generic.go:334] "Generic (PLEG): container finished" podID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" exitCode=0 Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834609 4960 generic.go:334] "Generic (PLEG): container finished" podID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" exitCode=0 Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerDied","Data":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834632 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834654 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerDied","Data":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834680 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerDied","Data":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerDied","Data":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0349dc-1091-4e8f-921f-71a349b674d2","Type":"ContainerDied","Data":"6a4bb08a63a51053ddfd7a634f043805750555a07c6547224602c5380bbbb0c4"} Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.834712 4960 scope.go:117] "RemoveContainer" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.840052 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts" (OuterVolumeSpecName: "scripts") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.898295 4960 scope.go:117] "RemoveContainer" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.898422 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.925640 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0349dc-1091-4e8f-921f-71a349b674d2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.925674 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.925686 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gv2f\" (UniqueName: \"kubernetes.io/projected/cd0349dc-1091-4e8f-921f-71a349b674d2-kube-api-access-4gv2f\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.925697 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.929270 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.951255 4960 scope.go:117] "RemoveContainer" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.981233 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.986880 4960 scope.go:117] "RemoveContainer" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:18 crc kubenswrapper[4960]: I1201 20:03:18.990898 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data" (OuterVolumeSpecName: "config-data") pod "cd0349dc-1091-4e8f-921f-71a349b674d2" (UID: "cd0349dc-1091-4e8f-921f-71a349b674d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.021957 4960 scope.go:117] "RemoveContainer" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.022528 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": container with ID starting with a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455 not found: ID does not exist" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.022554 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} err="failed to get container status \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": rpc error: code = NotFound desc = could not find container \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": container with ID starting with a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.022578 4960 scope.go:117] "RemoveContainer" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.022950 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": container with ID starting with 93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021 not found: ID does not exist" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.022969 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} err="failed to get container status \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": rpc error: code = NotFound desc = could not find container \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": container with ID starting with 93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.022981 4960 scope.go:117] "RemoveContainer" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.023296 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": container with ID starting with e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8 not found: ID does not exist" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.023317 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} err="failed to get container status \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": rpc error: code = NotFound desc = could not find container \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": container with ID starting with e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.023330 4960 scope.go:117] "RemoveContainer" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.023657 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": container with ID starting with 14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e not found: ID does not exist" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.023702 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} err="failed to get container status \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": rpc error: code = NotFound desc = could not find container \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": container with ID starting with 14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.023731 4960 scope.go:117] "RemoveContainer" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024103 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} err="failed to get container status \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": rpc error: code = NotFound desc = could not find container \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": container with ID starting with a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024140 4960 scope.go:117] "RemoveContainer" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024396 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} err="failed to get container status \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": rpc error: code = NotFound desc = could not find container \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": container with ID starting with 93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024414 4960 scope.go:117] "RemoveContainer" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024789 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} err="failed to get container status \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": rpc error: code = NotFound desc = could not find container \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": container with ID starting with e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.024913 4960 scope.go:117] "RemoveContainer" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.025306 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} err="failed to get container status \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": rpc error: code = NotFound desc = could not find container \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": container with ID starting with 14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.025396 4960 scope.go:117] "RemoveContainer" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.026211 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} err="failed to get container status \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": rpc error: code = NotFound desc = could not find container \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": container with ID starting with a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.026304 4960 scope.go:117] "RemoveContainer" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.026672 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} err="failed to get container status \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": rpc error: code = NotFound desc = could not find container \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": container with ID starting with 93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.026695 4960 scope.go:117] "RemoveContainer" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.027272 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} err="failed to get container status \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": rpc error: code = NotFound desc = could not find container \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": container with ID starting with e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.027295 4960 scope.go:117] "RemoveContainer" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.027445 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.027477 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0349dc-1091-4e8f-921f-71a349b674d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.028294 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} err="failed to get container status \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": rpc error: code = NotFound desc = could not find container \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": container with ID starting with 14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.028317 4960 scope.go:117] "RemoveContainer" containerID="a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.028625 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455"} err="failed to get container status \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": rpc error: code = NotFound desc = could not find container \"a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455\": container with ID starting with a25cd6c9ebed71554a82a69ec7d0f6d1fa8379513d31a97207bdaefbff13a455 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.028649 4960 scope.go:117] "RemoveContainer" containerID="93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.028914 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021"} err="failed to get container status \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": rpc error: code = NotFound desc = could not find container \"93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021\": container with ID starting with 93dd87b48c8ca65ab95517d9e362bb79f2f4a9a1a734da9ba235d3d6c2dbe021 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.029005 4960 scope.go:117] "RemoveContainer" containerID="e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.029341 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8"} err="failed to get container status \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": rpc error: code = NotFound desc = could not find container \"e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8\": container with ID starting with e1e0f95233d22a1b865641296eb59025056e4ef370f9037d8ec6ff26ca2f29d8 not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.029438 4960 scope.go:117] "RemoveContainer" containerID="14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.029715 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e"} err="failed to get container status \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": rpc error: code = NotFound desc = could not find container \"14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e\": container with ID starting with 14659a447c1bd76fbcb6fec0fdcf0175d75d7ccf2a653f749782156a38cdb08e not found: ID does not exist" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.184677 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.197163 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.208785 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.209307 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-central-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.209323 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-central-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.209361 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="sg-core" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.209367 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="sg-core" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.209376 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-notification-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.209384 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-notification-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: E1201 20:03:19.209395 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="proxy-httpd" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.209400 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="proxy-httpd" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.210004 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="proxy-httpd" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.210026 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="sg-core" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.210041 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-central-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.210057 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" containerName="ceilometer-notification-agent" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.211782 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.219638 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.219787 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.291575 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334319 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334423 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334460 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd6tb\" (UniqueName: \"kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334620 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334690 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.334720 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.343939 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ed64227-884f-41f4-8534-1bc160d45726" path="/var/lib/kubelet/pods/3ed64227-884f-41f4-8534-1bc160d45726/volumes" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.346924 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd0349dc-1091-4e8f-921f-71a349b674d2" path="/var/lib/kubelet/pods/cd0349dc-1091-4e8f-921f-71a349b674d2/volumes" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.437374 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd6tb\" (UniqueName: \"kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.437867 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.437969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.438036 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.438089 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.438224 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.438265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.439982 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.440237 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.443063 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.443620 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.445058 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.451618 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.454048 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd6tb\" (UniqueName: \"kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb\") pod \"ceilometer-0\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.608676 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.894061 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dknn6"] Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.895748 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.898716 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.899201 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.899347 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-blnkc" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.921134 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a341d11-27da-4def-a478-c878a69e2e6b","Type":"ContainerStarted","Data":"deba07a5aaec9aa2b483f98ec61c9f6306f38970b29c076da752870d1ac10d27"} Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.956142 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dknn6"] Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.960385 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerDied","Data":"def26066f8c718e9c7360c95a45942c4818705caa17fa1ccc96fee88aec073f7"} Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.959998 4960 generic.go:334] "Generic (PLEG): container finished" podID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerID="def26066f8c718e9c7360c95a45942c4818705caa17fa1ccc96fee88aec073f7" exitCode=0 Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.956271 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.963246 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.963322 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fzfr\" (UniqueName: \"kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.964463 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:19 crc kubenswrapper[4960]: I1201 20:03:19.992377 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065274 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065334 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065353 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065625 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065763 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065802 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.065830 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6hdq\" (UniqueName: \"kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq\") pod \"37988728-7b89-4f2c-bb2d-979a5339be9e\" (UID: \"37988728-7b89-4f2c-bb2d-979a5339be9e\") " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.066094 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.066163 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.066248 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.066278 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fzfr\" (UniqueName: \"kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.067057 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.069062 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs" (OuterVolumeSpecName: "logs") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.075748 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq" (OuterVolumeSpecName: "kube-api-access-n6hdq") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "kube-api-access-n6hdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.076816 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts" (OuterVolumeSpecName: "scripts") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.077433 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.085109 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fzfr\" (UniqueName: \"kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.095897 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.107373 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-dknn6\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.109218 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0" (OuterVolumeSpecName: "glance") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.137743 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.161543 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168562 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") on node \"crc\" " Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168602 4960 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168614 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37988728-7b89-4f2c-bb2d-979a5339be9e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168624 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6hdq\" (UniqueName: \"kubernetes.io/projected/37988728-7b89-4f2c-bb2d-979a5339be9e-kube-api-access-n6hdq\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168643 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.168651 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.199745 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.199892 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0") on node "crc" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.218354 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data" (OuterVolumeSpecName: "config-data") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.233381 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "37988728-7b89-4f2c-bb2d-979a5339be9e" (UID: "37988728-7b89-4f2c-bb2d-979a5339be9e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.270361 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.271952 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.272005 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.272026 4960 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/37988728-7b89-4f2c-bb2d-979a5339be9e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.322969 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.772797 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dknn6"] Dec 01 20:03:20 crc kubenswrapper[4960]: W1201 20:03:20.776382 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0048f864_a571_46d7_9f02_ec267bd88790.slice/crio-716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba WatchSource:0}: Error finding container 716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba: Status 404 returned error can't find the container with id 716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.977911 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerStarted","Data":"bde59f4bdee8e6809760fb4347c71f898949ddb7f8a2a60f950c5ac3456a8acf"} Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.979583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dknn6" event={"ID":"0048f864-a571-46d7-9f02-ec267bd88790","Type":"ContainerStarted","Data":"716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba"} Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.981907 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a341d11-27da-4def-a478-c878a69e2e6b","Type":"ContainerStarted","Data":"86371225b8ea38bb23427373c4fdac6c5800959a7c3137c0e84170e8475bc577"} Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.981949 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a341d11-27da-4def-a478-c878a69e2e6b","Type":"ContainerStarted","Data":"a3029dd0b0c90e79844a946eaa59f09001b077d3da2720f17f81a6042853193f"} Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.984070 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"37988728-7b89-4f2c-bb2d-979a5339be9e","Type":"ContainerDied","Data":"ee01c02c12f398f06b8b1904767bd31aa41642e39b25b4f84463bf75c3b92923"} Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.984155 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:20 crc kubenswrapper[4960]: I1201 20:03:20.984161 4960 scope.go:117] "RemoveContainer" containerID="def26066f8c718e9c7360c95a45942c4818705caa17fa1ccc96fee88aec073f7" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.021457 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.021435239 podStartE2EDuration="4.021435239s" podCreationTimestamp="2025-12-01 20:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:21.01122274 +0000 UTC m=+1436.298714419" watchObservedRunningTime="2025-12-01 20:03:21.021435239 +0000 UTC m=+1436.308926908" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.026373 4960 scope.go:117] "RemoveContainer" containerID="8fd8e20abe45a1b5aa11c7209951adfdad2e47d20e9cb8f61bb109149a87f0b7" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.035737 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.045230 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.054566 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:21 crc kubenswrapper[4960]: E1201 20:03:21.054972 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-log" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.054997 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-log" Dec 01 20:03:21 crc kubenswrapper[4960]: E1201 20:03:21.055021 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-httpd" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.055028 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-httpd" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.055231 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-httpd" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.055256 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" containerName="glance-log" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.056446 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.058996 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.059212 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.072586 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.195822 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s2k6\" (UniqueName: \"kubernetes.io/projected/a1f3729c-7da0-4c16-b370-cacabf55f2b7-kube-api-access-7s2k6\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196202 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196235 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196296 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196323 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196357 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.196393 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.297877 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.297947 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.297990 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298049 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s2k6\" (UniqueName: \"kubernetes.io/projected/a1f3729c-7da0-4c16-b370-cacabf55f2b7-kube-api-access-7s2k6\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298070 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298102 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298159 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298227 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298417 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.298545 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1f3729c-7da0-4c16-b370-cacabf55f2b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.306980 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.307020 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.307052 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.307368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1f3729c-7da0-4c16-b370-cacabf55f2b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.308617 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.308646 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/720c6baa92eb97844a8354a6d35963468f1db26ad29581ee9ba5e8c56ff873e5/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.314636 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s2k6\" (UniqueName: \"kubernetes.io/projected/a1f3729c-7da0-4c16-b370-cacabf55f2b7-kube-api-access-7s2k6\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.353879 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37988728-7b89-4f2c-bb2d-979a5339be9e" path="/var/lib/kubelet/pods/37988728-7b89-4f2c-bb2d-979a5339be9e/volumes" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.365777 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-faa3aa8a-0c48-4bfd-9625-e9bc554e8fb0\") pod \"glance-default-internal-api-0\" (UID: \"a1f3729c-7da0-4c16-b370-cacabf55f2b7\") " pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.382104 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.970305 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 01 20:03:21 crc kubenswrapper[4960]: I1201 20:03:21.997004 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1f3729c-7da0-4c16-b370-cacabf55f2b7","Type":"ContainerStarted","Data":"11bd774a971f60d987ea15cd10c7eeb8ae3d7b6096b7513e0b4bf4de8fc23427"} Dec 01 20:03:22 crc kubenswrapper[4960]: I1201 20:03:22.005220 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerStarted","Data":"24071666d34ea40ae3f4dcd1f4a11f31ae49de921f109f473a2b45ef81a1e362"} Dec 01 20:03:23 crc kubenswrapper[4960]: I1201 20:03:23.043814 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1f3729c-7da0-4c16-b370-cacabf55f2b7","Type":"ContainerStarted","Data":"370f8e18b45bc9a2a261ff7347409c2a9892a07a6acce252e41d92ac3a4b06b0"} Dec 01 20:03:23 crc kubenswrapper[4960]: I1201 20:03:23.047631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerStarted","Data":"b3b777be2cc0e7a77721103a5ef05bf864bb9a5efc7691cf2cf9d60a89446a1d"} Dec 01 20:03:24 crc kubenswrapper[4960]: I1201 20:03:24.076158 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a1f3729c-7da0-4c16-b370-cacabf55f2b7","Type":"ContainerStarted","Data":"7d1031b635b93c3b7debd4f4f214759b037e860c0d6eac55cd4a74fcf2b962f0"} Dec 01 20:03:24 crc kubenswrapper[4960]: I1201 20:03:24.086199 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerStarted","Data":"39346d8e19f34a4f3f89f8ff70bbf9cb362f0bfc94569499866782385d3606f0"} Dec 01 20:03:24 crc kubenswrapper[4960]: I1201 20:03:24.096788 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.096772978 podStartE2EDuration="3.096772978s" podCreationTimestamp="2025-12-01 20:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:24.091296627 +0000 UTC m=+1439.378788296" watchObservedRunningTime="2025-12-01 20:03:24.096772978 +0000 UTC m=+1439.384264647" Dec 01 20:03:24 crc kubenswrapper[4960]: I1201 20:03:24.564276 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107251 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerStarted","Data":"471874151c3cae42c895819fdc65df995906da6c637dd6c1f96e36cc4c6d2dcd"} Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107341 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-central-agent" containerID="cri-o://24071666d34ea40ae3f4dcd1f4a11f31ae49de921f109f473a2b45ef81a1e362" gracePeriod=30 Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107401 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="proxy-httpd" containerID="cri-o://471874151c3cae42c895819fdc65df995906da6c637dd6c1f96e36cc4c6d2dcd" gracePeriod=30 Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107460 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-notification-agent" containerID="cri-o://b3b777be2cc0e7a77721103a5ef05bf864bb9a5efc7691cf2cf9d60a89446a1d" gracePeriod=30 Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107473 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="sg-core" containerID="cri-o://39346d8e19f34a4f3f89f8ff70bbf9cb362f0bfc94569499866782385d3606f0" gracePeriod=30 Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.107920 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:03:25 crc kubenswrapper[4960]: I1201 20:03:25.133486 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.787784869 podStartE2EDuration="6.133467158s" podCreationTimestamp="2025-12-01 20:03:19 +0000 UTC" firstStartedPulling="2025-12-01 20:03:20.347573634 +0000 UTC m=+1435.635065303" lastFinishedPulling="2025-12-01 20:03:24.693255923 +0000 UTC m=+1439.980747592" observedRunningTime="2025-12-01 20:03:25.132355952 +0000 UTC m=+1440.419847621" watchObservedRunningTime="2025-12-01 20:03:25.133467158 +0000 UTC m=+1440.420958837" Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125520 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerDied","Data":"471874151c3cae42c895819fdc65df995906da6c637dd6c1f96e36cc4c6d2dcd"} Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125673 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerID="471874151c3cae42c895819fdc65df995906da6c637dd6c1f96e36cc4c6d2dcd" exitCode=0 Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125702 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerID="39346d8e19f34a4f3f89f8ff70bbf9cb362f0bfc94569499866782385d3606f0" exitCode=2 Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125709 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerID="b3b777be2cc0e7a77721103a5ef05bf864bb9a5efc7691cf2cf9d60a89446a1d" exitCode=0 Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125731 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerDied","Data":"39346d8e19f34a4f3f89f8ff70bbf9cb362f0bfc94569499866782385d3606f0"} Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.125759 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerDied","Data":"b3b777be2cc0e7a77721103a5ef05bf864bb9a5efc7691cf2cf9d60a89446a1d"} Dec 01 20:03:26 crc kubenswrapper[4960]: I1201 20:03:26.638251 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 01 20:03:28 crc kubenswrapper[4960]: I1201 20:03:28.322607 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 20:03:28 crc kubenswrapper[4960]: I1201 20:03:28.322899 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 01 20:03:28 crc kubenswrapper[4960]: I1201 20:03:28.364501 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 20:03:28 crc kubenswrapper[4960]: I1201 20:03:28.377593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 01 20:03:29 crc kubenswrapper[4960]: I1201 20:03:29.164365 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 20:03:29 crc kubenswrapper[4960]: I1201 20:03:29.164745 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.195785 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.197889 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.197891 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dknn6" event={"ID":"0048f864-a571-46d7-9f02-ec267bd88790","Type":"ContainerStarted","Data":"f3c7043a62c3cb27566e56f287428770bb3c204811b417c181879cc68a3e84ec"} Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.198606 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.231687 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-dknn6" podStartSLOduration=2.176292045 podStartE2EDuration="12.231668703s" podCreationTimestamp="2025-12-01 20:03:19 +0000 UTC" firstStartedPulling="2025-12-01 20:03:20.779236081 +0000 UTC m=+1436.066727750" lastFinishedPulling="2025-12-01 20:03:30.834612739 +0000 UTC m=+1446.122104408" observedRunningTime="2025-12-01 20:03:31.227686429 +0000 UTC m=+1446.515178088" watchObservedRunningTime="2025-12-01 20:03:31.231668703 +0000 UTC m=+1446.519160372" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.383195 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.383244 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.424752 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:31 crc kubenswrapper[4960]: I1201 20:03:31.436082 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:32 crc kubenswrapper[4960]: I1201 20:03:32.208821 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerID="24071666d34ea40ae3f4dcd1f4a11f31ae49de921f109f473a2b45ef81a1e362" exitCode=0 Dec 01 20:03:32 crc kubenswrapper[4960]: I1201 20:03:32.208966 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerDied","Data":"24071666d34ea40ae3f4dcd1f4a11f31ae49de921f109f473a2b45ef81a1e362"} Dec 01 20:03:32 crc kubenswrapper[4960]: I1201 20:03:32.209589 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:32 crc kubenswrapper[4960]: I1201 20:03:32.209623 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.840862 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972458 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972515 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972559 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972677 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972737 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd6tb\" (UniqueName: \"kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.972832 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd\") pod \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\" (UID: \"4c5ca550-e0ea-46bd-ba8f-8a60165b3119\") " Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.973361 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.973610 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.982660 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb" (OuterVolumeSpecName: "kube-api-access-bd6tb") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "kube-api-access-bd6tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:33 crc kubenswrapper[4960]: I1201 20:03:33.984295 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts" (OuterVolumeSpecName: "scripts") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.008273 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.058894 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075880 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075913 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075923 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075937 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd6tb\" (UniqueName: \"kubernetes.io/projected/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-kube-api-access-bd6tb\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075947 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.075955 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.091558 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data" (OuterVolumeSpecName: "config-data") pod "4c5ca550-e0ea-46bd-ba8f-8a60165b3119" (UID: "4c5ca550-e0ea-46bd-ba8f-8a60165b3119"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.177762 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c5ca550-e0ea-46bd-ba8f-8a60165b3119-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.232419 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c5ca550-e0ea-46bd-ba8f-8a60165b3119","Type":"ContainerDied","Data":"bde59f4bdee8e6809760fb4347c71f898949ddb7f8a2a60f950c5ac3456a8acf"} Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.232483 4960 scope.go:117] "RemoveContainer" containerID="471874151c3cae42c895819fdc65df995906da6c637dd6c1f96e36cc4c6d2dcd" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.232511 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.274755 4960 scope.go:117] "RemoveContainer" containerID="39346d8e19f34a4f3f89f8ff70bbf9cb362f0bfc94569499866782385d3606f0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.293156 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.303176 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.313878 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:34 crc kubenswrapper[4960]: E1201 20:03:34.318679 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="proxy-httpd" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.318713 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="proxy-httpd" Dec 01 20:03:34 crc kubenswrapper[4960]: E1201 20:03:34.318730 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-central-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.318740 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-central-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: E1201 20:03:34.318756 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="sg-core" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.318762 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="sg-core" Dec 01 20:03:34 crc kubenswrapper[4960]: E1201 20:03:34.318771 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-notification-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.318779 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-notification-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.318968 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-notification-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.319010 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="sg-core" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.319017 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="ceilometer-central-agent" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.319028 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" containerName="proxy-httpd" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.320943 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.324288 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.329251 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.329731 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.336665 4960 scope.go:117] "RemoveContainer" containerID="b3b777be2cc0e7a77721103a5ef05bf864bb9a5efc7691cf2cf9d60a89446a1d" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.364599 4960 scope.go:117] "RemoveContainer" containerID="24071666d34ea40ae3f4dcd1f4a11f31ae49de921f109f473a2b45ef81a1e362" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.382651 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.382740 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.382843 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.382915 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6djmm\" (UniqueName: \"kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.383012 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.383041 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.383354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.418593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.418677 4960 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485650 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485697 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485797 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485881 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485917 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485959 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.485986 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6djmm\" (UniqueName: \"kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.486589 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.486675 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.493273 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.493506 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.493788 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.493880 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.505720 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6djmm\" (UniqueName: \"kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm\") pod \"ceilometer-0\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " pod="openstack/ceilometer-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.596189 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 01 20:03:34 crc kubenswrapper[4960]: I1201 20:03:34.662767 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:03:35 crc kubenswrapper[4960]: W1201 20:03:35.198842 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c122f7_b0b2_4989_9984_3c3859eda76f.slice/crio-00854cdc42cf48f4c8b2ec625df0bae2764b9005327587a715e01263b0584233 WatchSource:0}: Error finding container 00854cdc42cf48f4c8b2ec625df0bae2764b9005327587a715e01263b0584233: Status 404 returned error can't find the container with id 00854cdc42cf48f4c8b2ec625df0bae2764b9005327587a715e01263b0584233 Dec 01 20:03:35 crc kubenswrapper[4960]: I1201 20:03:35.199961 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:03:35 crc kubenswrapper[4960]: I1201 20:03:35.247193 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerStarted","Data":"00854cdc42cf48f4c8b2ec625df0bae2764b9005327587a715e01263b0584233"} Dec 01 20:03:35 crc kubenswrapper[4960]: I1201 20:03:35.340351 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5ca550-e0ea-46bd-ba8f-8a60165b3119" path="/var/lib/kubelet/pods/4c5ca550-e0ea-46bd-ba8f-8a60165b3119/volumes" Dec 01 20:03:36 crc kubenswrapper[4960]: I1201 20:03:36.261281 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerStarted","Data":"12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e"} Dec 01 20:03:37 crc kubenswrapper[4960]: I1201 20:03:37.273256 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerStarted","Data":"193dbfa7c1f3250c28521d277ce6db04be87cc53c5ec0411e2380b5c445da478"} Dec 01 20:03:38 crc kubenswrapper[4960]: I1201 20:03:38.289303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerStarted","Data":"f86f8c0e3a4917ef1b4054f6d1150f6ea1c550e8f4787de54b2716f1649162cb"} Dec 01 20:03:40 crc kubenswrapper[4960]: I1201 20:03:40.311264 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerStarted","Data":"5bbf6281627287b92107ee9b051ae662a9d1aa9c0376dc47e7485aff04077593"} Dec 01 20:03:40 crc kubenswrapper[4960]: I1201 20:03:40.312308 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:03:40 crc kubenswrapper[4960]: I1201 20:03:40.345311 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.520819484 podStartE2EDuration="6.345288634s" podCreationTimestamp="2025-12-01 20:03:34 +0000 UTC" firstStartedPulling="2025-12-01 20:03:35.203290758 +0000 UTC m=+1450.490782427" lastFinishedPulling="2025-12-01 20:03:39.027759868 +0000 UTC m=+1454.315251577" observedRunningTime="2025-12-01 20:03:40.339276805 +0000 UTC m=+1455.626768484" watchObservedRunningTime="2025-12-01 20:03:40.345288634 +0000 UTC m=+1455.632780303" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.366715 4960 generic.go:334] "Generic (PLEG): container finished" podID="0048f864-a571-46d7-9f02-ec267bd88790" containerID="f3c7043a62c3cb27566e56f287428770bb3c204811b417c181879cc68a3e84ec" exitCode=0 Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.366823 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dknn6" event={"ID":"0048f864-a571-46d7-9f02-ec267bd88790","Type":"ContainerDied","Data":"f3c7043a62c3cb27566e56f287428770bb3c204811b417c181879cc68a3e84ec"} Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.403639 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.413679 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.436142 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.483451 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdwxq\" (UniqueName: \"kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.483727 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.484080 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.585857 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.585939 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdwxq\" (UniqueName: \"kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.586020 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.586481 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.586514 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.611467 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdwxq\" (UniqueName: \"kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq\") pod \"redhat-marketplace-c45pn\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.744056 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.872393 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:03:43 crc kubenswrapper[4960]: I1201 20:03:43.872471 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:03:44 crc kubenswrapper[4960]: W1201 20:03:44.247896 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97abc88d_b242_469e_82de_8cc4640b3761.slice/crio-adae3c8eb356155a21b11f000f7c54e19a1856950602f83dde27caae43f66160 WatchSource:0}: Error finding container adae3c8eb356155a21b11f000f7c54e19a1856950602f83dde27caae43f66160: Status 404 returned error can't find the container with id adae3c8eb356155a21b11f000f7c54e19a1856950602f83dde27caae43f66160 Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.248132 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.376735 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerStarted","Data":"adae3c8eb356155a21b11f000f7c54e19a1856950602f83dde27caae43f66160"} Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.807977 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.908544 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fzfr\" (UniqueName: \"kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr\") pod \"0048f864-a571-46d7-9f02-ec267bd88790\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.908657 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data\") pod \"0048f864-a571-46d7-9f02-ec267bd88790\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.908758 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts\") pod \"0048f864-a571-46d7-9f02-ec267bd88790\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.908791 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle\") pod \"0048f864-a571-46d7-9f02-ec267bd88790\" (UID: \"0048f864-a571-46d7-9f02-ec267bd88790\") " Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.915586 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts" (OuterVolumeSpecName: "scripts") pod "0048f864-a571-46d7-9f02-ec267bd88790" (UID: "0048f864-a571-46d7-9f02-ec267bd88790"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.915638 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr" (OuterVolumeSpecName: "kube-api-access-6fzfr") pod "0048f864-a571-46d7-9f02-ec267bd88790" (UID: "0048f864-a571-46d7-9f02-ec267bd88790"). InnerVolumeSpecName "kube-api-access-6fzfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.938919 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0048f864-a571-46d7-9f02-ec267bd88790" (UID: "0048f864-a571-46d7-9f02-ec267bd88790"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:44 crc kubenswrapper[4960]: I1201 20:03:44.958215 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data" (OuterVolumeSpecName: "config-data") pod "0048f864-a571-46d7-9f02-ec267bd88790" (UID: "0048f864-a571-46d7-9f02-ec267bd88790"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.016576 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.016622 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.016640 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0048f864-a571-46d7-9f02-ec267bd88790-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.016661 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fzfr\" (UniqueName: \"kubernetes.io/projected/0048f864-a571-46d7-9f02-ec267bd88790-kube-api-access-6fzfr\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.391222 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-dknn6" event={"ID":"0048f864-a571-46d7-9f02-ec267bd88790","Type":"ContainerDied","Data":"716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba"} Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.391498 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="716d4697935d0ef260ad03bae4f4b6a3dd5cf3782c98ac2e494efd9ced4088ba" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.391290 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-dknn6" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.393543 4960 generic.go:334] "Generic (PLEG): container finished" podID="97abc88d-b242-469e-82de-8cc4640b3761" containerID="93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b" exitCode=0 Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.393607 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerDied","Data":"93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b"} Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.537837 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 20:03:45 crc kubenswrapper[4960]: E1201 20:03:45.538333 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0048f864-a571-46d7-9f02-ec267bd88790" containerName="nova-cell0-conductor-db-sync" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.538350 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0048f864-a571-46d7-9f02-ec267bd88790" containerName="nova-cell0-conductor-db-sync" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.538544 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0048f864-a571-46d7-9f02-ec267bd88790" containerName="nova-cell0-conductor-db-sync" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.539347 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.542086 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.542542 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-blnkc" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.562497 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.631032 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.631099 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q99jl\" (UniqueName: \"kubernetes.io/projected/878b6439-3b49-40d4-80e2-f01694590b67-kube-api-access-q99jl\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.631234 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.733017 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.733168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.733213 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q99jl\" (UniqueName: \"kubernetes.io/projected/878b6439-3b49-40d4-80e2-f01694590b67-kube-api-access-q99jl\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.737936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.738197 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b6439-3b49-40d4-80e2-f01694590b67-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.751011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q99jl\" (UniqueName: \"kubernetes.io/projected/878b6439-3b49-40d4-80e2-f01694590b67-kube-api-access-q99jl\") pod \"nova-cell0-conductor-0\" (UID: \"878b6439-3b49-40d4-80e2-f01694590b67\") " pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:45 crc kubenswrapper[4960]: I1201 20:03:45.864824 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:46 crc kubenswrapper[4960]: W1201 20:03:46.357009 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod878b6439_3b49_40d4_80e2_f01694590b67.slice/crio-404d998fb598501fe25ee36a6cb7768b3521368c7e97a76bd11dd17114ae3d64 WatchSource:0}: Error finding container 404d998fb598501fe25ee36a6cb7768b3521368c7e97a76bd11dd17114ae3d64: Status 404 returned error can't find the container with id 404d998fb598501fe25ee36a6cb7768b3521368c7e97a76bd11dd17114ae3d64 Dec 01 20:03:46 crc kubenswrapper[4960]: I1201 20:03:46.361442 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 01 20:03:46 crc kubenswrapper[4960]: I1201 20:03:46.403654 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"878b6439-3b49-40d4-80e2-f01694590b67","Type":"ContainerStarted","Data":"404d998fb598501fe25ee36a6cb7768b3521368c7e97a76bd11dd17114ae3d64"} Dec 01 20:03:47 crc kubenswrapper[4960]: I1201 20:03:47.420031 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"878b6439-3b49-40d4-80e2-f01694590b67","Type":"ContainerStarted","Data":"49ccc43beb4a3adf0ecbfee3ddd9539c83029358b2a68dd807a50dd6981e07b0"} Dec 01 20:03:47 crc kubenswrapper[4960]: I1201 20:03:47.420471 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:48 crc kubenswrapper[4960]: I1201 20:03:48.434024 4960 generic.go:334] "Generic (PLEG): container finished" podID="97abc88d-b242-469e-82de-8cc4640b3761" containerID="a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c" exitCode=0 Dec 01 20:03:48 crc kubenswrapper[4960]: I1201 20:03:48.434098 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerDied","Data":"a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c"} Dec 01 20:03:48 crc kubenswrapper[4960]: I1201 20:03:48.469676 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.46965614 podStartE2EDuration="3.46965614s" podCreationTimestamp="2025-12-01 20:03:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:47.44155915 +0000 UTC m=+1462.729050819" watchObservedRunningTime="2025-12-01 20:03:48.46965614 +0000 UTC m=+1463.757147819" Dec 01 20:03:49 crc kubenswrapper[4960]: I1201 20:03:49.448718 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerStarted","Data":"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d"} Dec 01 20:03:49 crc kubenswrapper[4960]: I1201 20:03:49.478035 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c45pn" podStartSLOduration=2.882214308 podStartE2EDuration="6.478015492s" podCreationTimestamp="2025-12-01 20:03:43 +0000 UTC" firstStartedPulling="2025-12-01 20:03:45.396850491 +0000 UTC m=+1460.684342170" lastFinishedPulling="2025-12-01 20:03:48.992651665 +0000 UTC m=+1464.280143354" observedRunningTime="2025-12-01 20:03:49.467853004 +0000 UTC m=+1464.755344693" watchObservedRunningTime="2025-12-01 20:03:49.478015492 +0000 UTC m=+1464.765507161" Dec 01 20:03:53 crc kubenswrapper[4960]: I1201 20:03:53.745443 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:53 crc kubenswrapper[4960]: I1201 20:03:53.746172 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:53 crc kubenswrapper[4960]: I1201 20:03:53.808688 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:54 crc kubenswrapper[4960]: I1201 20:03:54.557921 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:54 crc kubenswrapper[4960]: I1201 20:03:54.614355 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:55 crc kubenswrapper[4960]: I1201 20:03:55.906862 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.804728 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c45pn" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="registry-server" containerID="cri-o://57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d" gracePeriod=2 Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.979650 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-8mhss"] Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.981947 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.986295 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.986309 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 01 20:03:56 crc kubenswrapper[4960]: I1201 20:03:56.999815 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8mhss"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.009405 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.009528 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.009594 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdgd2\" (UniqueName: \"kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.009668 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.112592 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.113002 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdgd2\" (UniqueName: \"kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.113066 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.113208 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.122391 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.128036 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.133058 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.139860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdgd2\" (UniqueName: \"kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2\") pod \"nova-cell0-cell-mapping-8mhss\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.197005 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.198334 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.204746 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.217179 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.273421 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.275475 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.281715 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.304700 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.313843 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.316295 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.316352 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdk8m\" (UniqueName: \"kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.316398 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6qxn\" (UniqueName: \"kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.316436 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.316480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.320835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.320887 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.396086 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.397917 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.406760 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.408381 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.411101 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423408 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6qxn\" (UniqueName: \"kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423457 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423507 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423539 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423556 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4hf\" (UniqueName: \"kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423613 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423631 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423657 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423709 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423730 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423753 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423837 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.423887 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdk8m\" (UniqueName: \"kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.428605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.429945 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.441629 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.453670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.471454 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.485005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdk8m\" (UniqueName: \"kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m\") pod \"nova-scheduler-0\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.485922 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.485933 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6qxn\" (UniqueName: \"kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn\") pod \"nova-metadata-0\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.501345 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.525431 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b776b\" (UniqueName: \"kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527093 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527269 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527393 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527471 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4hf\" (UniqueName: \"kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527588 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527678 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.527872 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.529597 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.529620 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.530355 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.530940 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.533381 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.534304 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.535161 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.537027 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.565313 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4hf\" (UniqueName: \"kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf\") pod \"dnsmasq-dns-78cd565959-6tjw9\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.565370 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.578874 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629591 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b776b\" (UniqueName: \"kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629642 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629663 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629689 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629717 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629834 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.629885 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjt5m\" (UniqueName: \"kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.636551 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.640294 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.647546 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.648205 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b776b\" (UniqueName: \"kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.659151 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.737522 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities\") pod \"97abc88d-b242-469e-82de-8cc4640b3761\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.737650 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdwxq\" (UniqueName: \"kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq\") pod \"97abc88d-b242-469e-82de-8cc4640b3761\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.737727 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content\") pod \"97abc88d-b242-469e-82de-8cc4640b3761\" (UID: \"97abc88d-b242-469e-82de-8cc4640b3761\") " Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.738024 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.738050 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.738090 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.738242 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjt5m\" (UniqueName: \"kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.738937 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities" (OuterVolumeSpecName: "utilities") pod "97abc88d-b242-469e-82de-8cc4640b3761" (UID: "97abc88d-b242-469e-82de-8cc4640b3761"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.739613 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.749382 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.755842 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.759124 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97abc88d-b242-469e-82de-8cc4640b3761" (UID: "97abc88d-b242-469e-82de-8cc4640b3761"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.762322 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq" (OuterVolumeSpecName: "kube-api-access-cdwxq") pod "97abc88d-b242-469e-82de-8cc4640b3761" (UID: "97abc88d-b242-469e-82de-8cc4640b3761"). InnerVolumeSpecName "kube-api-access-cdwxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.762974 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjt5m\" (UniqueName: \"kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m\") pod \"nova-api-0\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " pod="openstack/nova-api-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.826021 4960 generic.go:334] "Generic (PLEG): container finished" podID="97abc88d-b242-469e-82de-8cc4640b3761" containerID="57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d" exitCode=0 Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.826087 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerDied","Data":"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d"} Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.826150 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c45pn" event={"ID":"97abc88d-b242-469e-82de-8cc4640b3761","Type":"ContainerDied","Data":"adae3c8eb356155a21b11f000f7c54e19a1856950602f83dde27caae43f66160"} Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.826169 4960 scope.go:117] "RemoveContainer" containerID="57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.826240 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c45pn" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.839764 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.839790 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdwxq\" (UniqueName: \"kubernetes.io/projected/97abc88d-b242-469e-82de-8cc4640b3761-kube-api-access-cdwxq\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.839802 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97abc88d-b242-469e-82de-8cc4640b3761-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.851628 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.871219 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.873770 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.878950 4960 scope.go:117] "RemoveContainer" containerID="a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c" Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.889295 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c45pn"] Dec 01 20:03:57 crc kubenswrapper[4960]: I1201 20:03:57.892239 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.051936 4960 scope.go:117] "RemoveContainer" containerID="93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.068244 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8mhss"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.174761 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvgdt"] Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.180843 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="extract-utilities" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.180890 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="extract-utilities" Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.180973 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="registry-server" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.180980 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="registry-server" Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.181016 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="extract-content" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.181028 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="extract-content" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.181472 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="97abc88d-b242-469e-82de-8cc4640b3761" containerName="registry-server" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.183923 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.188567 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.191222 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.191226 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.218933 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvgdt"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.257383 4960 scope.go:117] "RemoveContainer" containerID="57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d" Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.263505 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d\": container with ID starting with 57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d not found: ID does not exist" containerID="57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.263548 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d"} err="failed to get container status \"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d\": rpc error: code = NotFound desc = could not find container \"57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d\": container with ID starting with 57871d8a66d15cd4c4eb5d3e0632741bdf771675d8e60ca8edec9565bb2dcc2d not found: ID does not exist" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.263577 4960 scope.go:117] "RemoveContainer" containerID="a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c" Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.267043 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c\": container with ID starting with a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c not found: ID does not exist" containerID="a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.267086 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c"} err="failed to get container status \"a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c\": rpc error: code = NotFound desc = could not find container \"a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c\": container with ID starting with a61f3ec68baf9cdb1e291d87b4130e9f62e2f36237d482f78557b2b646e63a3c not found: ID does not exist" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.267132 4960 scope.go:117] "RemoveContainer" containerID="93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b" Dec 01 20:03:58 crc kubenswrapper[4960]: E1201 20:03:58.269259 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b\": container with ID starting with 93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b not found: ID does not exist" containerID="93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.269284 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b"} err="failed to get container status \"93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b\": rpc error: code = NotFound desc = could not find container \"93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b\": container with ID starting with 93fd3d95cbcb1f3003acc95154d85362c70385dd57f899333f0ea73bc6255a3b not found: ID does not exist" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.298021 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.361775 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.361937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.362020 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.362073 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llf9c\" (UniqueName: \"kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.464903 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.465955 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.466063 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llf9c\" (UniqueName: \"kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.466248 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.471982 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.472781 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.476670 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.486686 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llf9c\" (UniqueName: \"kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c\") pod \"nova-cell1-conductor-db-sync-tvgdt\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.542804 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:03:58 crc kubenswrapper[4960]: W1201 20:03:58.639325 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c393fd6_f1d0_4c53_8558_e909279da2de.slice/crio-7eb35947787275e964d03f31d4e3fa3b9cda0e6e28d165860c3f0f9132f40ca7 WatchSource:0}: Error finding container 7eb35947787275e964d03f31d4e3fa3b9cda0e6e28d165860c3f0f9132f40ca7: Status 404 returned error can't find the container with id 7eb35947787275e964d03f31d4e3fa3b9cda0e6e28d165860c3f0f9132f40ca7 Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.646146 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.773308 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.790148 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:03:58 crc kubenswrapper[4960]: W1201 20:03:58.790717 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04f34f8e_c023_476a_9a1e_f9c266fd1b86.slice/crio-fa39f48bcfcf4a97b398aeeb62b4fb76edb109ab27a6cae41441db66fd34174f WatchSource:0}: Error finding container fa39f48bcfcf4a97b398aeeb62b4fb76edb109ab27a6cae41441db66fd34174f: Status 404 returned error can't find the container with id fa39f48bcfcf4a97b398aeeb62b4fb76edb109ab27a6cae41441db66fd34174f Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.837381 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8mhss" event={"ID":"82af32f2-8438-4070-814e-bfad17fe42b9","Type":"ContainerStarted","Data":"80b33f81a5841f3ec676a4a386b16a7c5b56e34fa76b76bcda3766b726a439d1"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.837452 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8mhss" event={"ID":"82af32f2-8438-4070-814e-bfad17fe42b9","Type":"ContainerStarted","Data":"412d3a92f35cfac5d372fc6e0eceebe07df9593489991746a7241259ea433b78"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.840903 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5","Type":"ContainerStarted","Data":"f08334c2f87c54e65965a87163130db90706f890e0bc8f8d64176eef9144221d"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.849555 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b5567cc-82b1-456a-ada3-2e6141808bfb","Type":"ContainerStarted","Data":"e0f78c7c2e3576959b46393eeacab95f25cf36b51f47533ba75131064a110dc3"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.856442 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-8mhss" podStartSLOduration=2.856379637 podStartE2EDuration="2.856379637s" podCreationTimestamp="2025-12-01 20:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:58.853177607 +0000 UTC m=+1474.140669276" watchObservedRunningTime="2025-12-01 20:03:58.856379637 +0000 UTC m=+1474.143871316" Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.862343 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" event={"ID":"04f34f8e-c023-476a-9a1e-f9c266fd1b86","Type":"ContainerStarted","Data":"fa39f48bcfcf4a97b398aeeb62b4fb76edb109ab27a6cae41441db66fd34174f"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.864804 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerStarted","Data":"7eb35947787275e964d03f31d4e3fa3b9cda0e6e28d165860c3f0f9132f40ca7"} Dec 01 20:03:58 crc kubenswrapper[4960]: I1201 20:03:58.867651 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerStarted","Data":"1088a7b00e6db6b3d8acb74246a2fbcdff84cb491142e09b7d77df1f38344cb3"} Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.048689 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvgdt"] Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.349806 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97abc88d-b242-469e-82de-8cc4640b3761" path="/var/lib/kubelet/pods/97abc88d-b242-469e-82de-8cc4640b3761/volumes" Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.884381 4960 generic.go:334] "Generic (PLEG): container finished" podID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerID="91f6896034eb54ed1ea0626a67f3fa5689f3c9ab41741b6e7db5788a56127b8c" exitCode=0 Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.884430 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" event={"ID":"04f34f8e-c023-476a-9a1e-f9c266fd1b86","Type":"ContainerDied","Data":"91f6896034eb54ed1ea0626a67f3fa5689f3c9ab41741b6e7db5788a56127b8c"} Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.887749 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" event={"ID":"5dae60dd-ebab-42c1-9dc6-722768140fc4","Type":"ContainerStarted","Data":"509edc4faa04faee93bd0e25d7b0404ded6224148eba909a1a0690edf60e7a26"} Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.887793 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" event={"ID":"5dae60dd-ebab-42c1-9dc6-722768140fc4","Type":"ContainerStarted","Data":"0eb522b116086926aca59012a63e8ec44a1619483c1e27409cd3e32aa949d969"} Dec 01 20:03:59 crc kubenswrapper[4960]: I1201 20:03:59.936844 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" podStartSLOduration=1.936821025 podStartE2EDuration="1.936821025s" podCreationTimestamp="2025-12-01 20:03:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:03:59.92773178 +0000 UTC m=+1475.215223449" watchObservedRunningTime="2025-12-01 20:03:59.936821025 +0000 UTC m=+1475.224312694" Dec 01 20:04:01 crc kubenswrapper[4960]: I1201 20:04:01.239084 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:01 crc kubenswrapper[4960]: I1201 20:04:01.248432 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.923249 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerStarted","Data":"516807a73e40f0edb3f2e5e1c2fed6d126ddb1a2e51b77f1f4fe8bc9a1d0ae33"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.923964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerStarted","Data":"a912d7be53949fe393f27f58542ae90ff94092398786345c76715fac25517052"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.923494 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-metadata" containerID="cri-o://516807a73e40f0edb3f2e5e1c2fed6d126ddb1a2e51b77f1f4fe8bc9a1d0ae33" gracePeriod=30 Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.923323 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-log" containerID="cri-o://a912d7be53949fe393f27f58542ae90ff94092398786345c76715fac25517052" gracePeriod=30 Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.928314 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0400687f51e97a3e36fbed612e1c6cc0fd75180255637a22ce2f9004419fdc9f" gracePeriod=30 Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.928349 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5","Type":"ContainerStarted","Data":"0400687f51e97a3e36fbed612e1c6cc0fd75180255637a22ce2f9004419fdc9f"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.931033 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b5567cc-82b1-456a-ada3-2e6141808bfb","Type":"ContainerStarted","Data":"8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.944434 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" event={"ID":"04f34f8e-c023-476a-9a1e-f9c266fd1b86","Type":"ContainerStarted","Data":"1e4d4c2ddcc5c90dd903e202b5a38b04cefabdaab87df392c4190da1dd2b0ef7"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.944574 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.944636 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.091187744 podStartE2EDuration="5.944611531s" podCreationTimestamp="2025-12-01 20:03:57 +0000 UTC" firstStartedPulling="2025-12-01 20:03:58.321538721 +0000 UTC m=+1473.609030390" lastFinishedPulling="2025-12-01 20:04:02.174962508 +0000 UTC m=+1477.462454177" observedRunningTime="2025-12-01 20:04:02.942202484 +0000 UTC m=+1478.229694163" watchObservedRunningTime="2025-12-01 20:04:02.944611531 +0000 UTC m=+1478.232103200" Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.949959 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerStarted","Data":"a0fd097755e2fabb955364cdcc48f93c04efbe40c5789108e55ca03ee0edc2cd"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.950022 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerStarted","Data":"f38da714308b5ddcb0bddb39b8781115a23ffdbafdf892341449fd1889d3410e"} Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.971332 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.055316681 podStartE2EDuration="5.971306505s" podCreationTimestamp="2025-12-01 20:03:57 +0000 UTC" firstStartedPulling="2025-12-01 20:03:58.257761066 +0000 UTC m=+1473.545252735" lastFinishedPulling="2025-12-01 20:04:02.17375089 +0000 UTC m=+1477.461242559" observedRunningTime="2025-12-01 20:04:02.962633034 +0000 UTC m=+1478.250124703" watchObservedRunningTime="2025-12-01 20:04:02.971306505 +0000 UTC m=+1478.258798174" Dec 01 20:04:02 crc kubenswrapper[4960]: I1201 20:04:02.978284 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.60143061 podStartE2EDuration="5.978259273s" podCreationTimestamp="2025-12-01 20:03:57 +0000 UTC" firstStartedPulling="2025-12-01 20:03:58.797069671 +0000 UTC m=+1474.084561330" lastFinishedPulling="2025-12-01 20:04:02.173898324 +0000 UTC m=+1477.461389993" observedRunningTime="2025-12-01 20:04:02.977956474 +0000 UTC m=+1478.265448143" watchObservedRunningTime="2025-12-01 20:04:02.978259273 +0000 UTC m=+1478.265750942" Dec 01 20:04:03 crc kubenswrapper[4960]: I1201 20:04:03.000631 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" podStartSLOduration=6.000616393 podStartE2EDuration="6.000616393s" podCreationTimestamp="2025-12-01 20:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:03.00051383 +0000 UTC m=+1478.288005499" watchObservedRunningTime="2025-12-01 20:04:03.000616393 +0000 UTC m=+1478.288108062" Dec 01 20:04:03 crc kubenswrapper[4960]: I1201 20:04:03.018407 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.485595715 podStartE2EDuration="6.018386118s" podCreationTimestamp="2025-12-01 20:03:57 +0000 UTC" firstStartedPulling="2025-12-01 20:03:58.642312408 +0000 UTC m=+1473.929804077" lastFinishedPulling="2025-12-01 20:04:02.175102811 +0000 UTC m=+1477.462594480" observedRunningTime="2025-12-01 20:04:03.015371684 +0000 UTC m=+1478.302863363" watchObservedRunningTime="2025-12-01 20:04:03.018386118 +0000 UTC m=+1478.305877787" Dec 01 20:04:03 crc kubenswrapper[4960]: I1201 20:04:03.969910 4960 generic.go:334] "Generic (PLEG): container finished" podID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerID="a912d7be53949fe393f27f58542ae90ff94092398786345c76715fac25517052" exitCode=143 Dec 01 20:04:03 crc kubenswrapper[4960]: I1201 20:04:03.970013 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerDied","Data":"a912d7be53949fe393f27f58542ae90ff94092398786345c76715fac25517052"} Dec 01 20:04:04 crc kubenswrapper[4960]: I1201 20:04:04.672280 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 20:04:05 crc kubenswrapper[4960]: I1201 20:04:05.995521 4960 generic.go:334] "Generic (PLEG): container finished" podID="82af32f2-8438-4070-814e-bfad17fe42b9" containerID="80b33f81a5841f3ec676a4a386b16a7c5b56e34fa76b76bcda3766b726a439d1" exitCode=0 Dec 01 20:04:05 crc kubenswrapper[4960]: I1201 20:04:05.995616 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8mhss" event={"ID":"82af32f2-8438-4070-814e-bfad17fe42b9","Type":"ContainerDied","Data":"80b33f81a5841f3ec676a4a386b16a7c5b56e34fa76b76bcda3766b726a439d1"} Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.009037 4960 generic.go:334] "Generic (PLEG): container finished" podID="5dae60dd-ebab-42c1-9dc6-722768140fc4" containerID="509edc4faa04faee93bd0e25d7b0404ded6224148eba909a1a0690edf60e7a26" exitCode=0 Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.009145 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" event={"ID":"5dae60dd-ebab-42c1-9dc6-722768140fc4","Type":"ContainerDied","Data":"509edc4faa04faee93bd0e25d7b0404ded6224148eba909a1a0690edf60e7a26"} Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.478674 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.579260 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.579424 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.598786 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts\") pod \"82af32f2-8438-4070-814e-bfad17fe42b9\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.598904 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdgd2\" (UniqueName: \"kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2\") pod \"82af32f2-8438-4070-814e-bfad17fe42b9\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.599028 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle\") pod \"82af32f2-8438-4070-814e-bfad17fe42b9\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.599208 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data\") pod \"82af32f2-8438-4070-814e-bfad17fe42b9\" (UID: \"82af32f2-8438-4070-814e-bfad17fe42b9\") " Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.607733 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts" (OuterVolumeSpecName: "scripts") pod "82af32f2-8438-4070-814e-bfad17fe42b9" (UID: "82af32f2-8438-4070-814e-bfad17fe42b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.610882 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2" (OuterVolumeSpecName: "kube-api-access-jdgd2") pod "82af32f2-8438-4070-814e-bfad17fe42b9" (UID: "82af32f2-8438-4070-814e-bfad17fe42b9"). InnerVolumeSpecName "kube-api-access-jdgd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.632252 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.642056 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82af32f2-8438-4070-814e-bfad17fe42b9" (UID: "82af32f2-8438-4070-814e-bfad17fe42b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.643473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data" (OuterVolumeSpecName: "config-data") pod "82af32f2-8438-4070-814e-bfad17fe42b9" (UID: "82af32f2-8438-4070-814e-bfad17fe42b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.648526 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.648599 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.701180 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdgd2\" (UniqueName: \"kubernetes.io/projected/82af32f2-8438-4070-814e-bfad17fe42b9-kube-api-access-jdgd2\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.701220 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.701234 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.701245 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82af32f2-8438-4070-814e-bfad17fe42b9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.853543 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.874412 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.892972 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.893050 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.927832 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:04:07 crc kubenswrapper[4960]: I1201 20:04:07.928033 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="dnsmasq-dns" containerID="cri-o://b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25" gracePeriod=10 Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.021787 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8mhss" event={"ID":"82af32f2-8438-4070-814e-bfad17fe42b9","Type":"ContainerDied","Data":"412d3a92f35cfac5d372fc6e0eceebe07df9593489991746a7241259ea433b78"} Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.022206 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="412d3a92f35cfac5d372fc6e0eceebe07df9593489991746a7241259ea433b78" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.022027 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8mhss" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.087026 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.187966 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.197406 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-log" containerID="cri-o://f38da714308b5ddcb0bddb39b8781115a23ffdbafdf892341449fd1889d3410e" gracePeriod=30 Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.197982 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-api" containerID="cri-o://a0fd097755e2fabb955364cdcc48f93c04efbe40c5789108e55ca03ee0edc2cd" gracePeriod=30 Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.204424 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.213:8774/\": EOF" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.204694 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.213:8774/\": EOF" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.541736 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.629900 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts\") pod \"5dae60dd-ebab-42c1-9dc6-722768140fc4\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.630504 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data\") pod \"5dae60dd-ebab-42c1-9dc6-722768140fc4\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.630540 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llf9c\" (UniqueName: \"kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c\") pod \"5dae60dd-ebab-42c1-9dc6-722768140fc4\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.630603 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle\") pod \"5dae60dd-ebab-42c1-9dc6-722768140fc4\" (UID: \"5dae60dd-ebab-42c1-9dc6-722768140fc4\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.638473 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c" (OuterVolumeSpecName: "kube-api-access-llf9c") pod "5dae60dd-ebab-42c1-9dc6-722768140fc4" (UID: "5dae60dd-ebab-42c1-9dc6-722768140fc4"). InnerVolumeSpecName "kube-api-access-llf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.660327 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts" (OuterVolumeSpecName: "scripts") pod "5dae60dd-ebab-42c1-9dc6-722768140fc4" (UID: "5dae60dd-ebab-42c1-9dc6-722768140fc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.680667 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.744123 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data" (OuterVolumeSpecName: "config-data") pod "5dae60dd-ebab-42c1-9dc6-722768140fc4" (UID: "5dae60dd-ebab-42c1-9dc6-722768140fc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.745372 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.745427 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.745441 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llf9c\" (UniqueName: \"kubernetes.io/projected/5dae60dd-ebab-42c1-9dc6-722768140fc4-kube-api-access-llf9c\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.759706 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5dae60dd-ebab-42c1-9dc6-722768140fc4" (UID: "5dae60dd-ebab-42c1-9dc6-722768140fc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.846493 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.846693 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.846854 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.846890 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbcxc\" (UniqueName: \"kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.847042 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.847092 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb\") pod \"18497e75-1fc4-4324-ab61-b067d19c1ce0\" (UID: \"18497e75-1fc4-4324-ab61-b067d19c1ce0\") " Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.847980 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dae60dd-ebab-42c1-9dc6-722768140fc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.851494 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc" (OuterVolumeSpecName: "kube-api-access-qbcxc") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "kube-api-access-qbcxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.910527 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.923495 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.937605 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.939878 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config" (OuterVolumeSpecName: "config") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.943403 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.950151 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.950193 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.950207 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbcxc\" (UniqueName: \"kubernetes.io/projected/18497e75-1fc4-4324-ab61-b067d19c1ce0-kube-api-access-qbcxc\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.950220 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.950230 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:08 crc kubenswrapper[4960]: I1201 20:04:08.951386 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18497e75-1fc4-4324-ab61-b067d19c1ce0" (UID: "18497e75-1fc4-4324-ab61-b067d19c1ce0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.032243 4960 generic.go:334] "Generic (PLEG): container finished" podID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerID="f38da714308b5ddcb0bddb39b8781115a23ffdbafdf892341449fd1889d3410e" exitCode=143 Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.032317 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerDied","Data":"f38da714308b5ddcb0bddb39b8781115a23ffdbafdf892341449fd1889d3410e"} Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.034141 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.034137 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tvgdt" event={"ID":"5dae60dd-ebab-42c1-9dc6-722768140fc4","Type":"ContainerDied","Data":"0eb522b116086926aca59012a63e8ec44a1619483c1e27409cd3e32aa949d969"} Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.034184 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eb522b116086926aca59012a63e8ec44a1619483c1e27409cd3e32aa949d969" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.035925 4960 generic.go:334] "Generic (PLEG): container finished" podID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerID="b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25" exitCode=0 Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.035978 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.036015 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerDied","Data":"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25"} Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.036066 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bdc55879-5cl7g" event={"ID":"18497e75-1fc4-4324-ab61-b067d19c1ce0","Type":"ContainerDied","Data":"0f2543b2077d3ab6df6492fde1518e407dd1246d1d90d7831a9e511a878b1fa6"} Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.036090 4960 scope.go:117] "RemoveContainer" containerID="b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.051967 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18497e75-1fc4-4324-ab61-b067d19c1ce0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.067513 4960 scope.go:117] "RemoveContainer" containerID="1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.089208 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.096664 4960 scope.go:117] "RemoveContainer" containerID="b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25" Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.097731 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25\": container with ID starting with b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25 not found: ID does not exist" containerID="b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.097784 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25"} err="failed to get container status \"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25\": rpc error: code = NotFound desc = could not find container \"b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25\": container with ID starting with b1d3e9b85014fdc05c64581045f8a73b2ab48493da3384235a688e2e48159f25 not found: ID does not exist" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.097814 4960 scope.go:117] "RemoveContainer" containerID="1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b" Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.098986 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b\": container with ID starting with 1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b not found: ID does not exist" containerID="1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.099036 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b"} err="failed to get container status \"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b\": rpc error: code = NotFound desc = could not find container \"1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b\": container with ID starting with 1f7b891d8bd0243daf94480c1a86fe06098b640a214a12bf9693d98037e16c0b not found: ID does not exist" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.102139 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bdc55879-5cl7g"] Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121003 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.121549 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="init" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121566 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="init" Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.121589 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82af32f2-8438-4070-814e-bfad17fe42b9" containerName="nova-manage" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121595 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="82af32f2-8438-4070-814e-bfad17fe42b9" containerName="nova-manage" Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.121614 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dae60dd-ebab-42c1-9dc6-722768140fc4" containerName="nova-cell1-conductor-db-sync" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121620 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dae60dd-ebab-42c1-9dc6-722768140fc4" containerName="nova-cell1-conductor-db-sync" Dec 01 20:04:09 crc kubenswrapper[4960]: E1201 20:04:09.121639 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="dnsmasq-dns" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121644 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="dnsmasq-dns" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121877 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" containerName="dnsmasq-dns" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121895 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="82af32f2-8438-4070-814e-bfad17fe42b9" containerName="nova-manage" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.121908 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dae60dd-ebab-42c1-9dc6-722768140fc4" containerName="nova-cell1-conductor-db-sync" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.122711 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.126271 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.136703 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.258191 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.258254 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swg5c\" (UniqueName: \"kubernetes.io/projected/38ad16c2-10a9-4a20-a070-e1ed80f62684-kube-api-access-swg5c\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.258304 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.268418 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.268618 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerName="kube-state-metrics" containerID="cri-o://c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650" gracePeriod=30 Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.342806 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18497e75-1fc4-4324-ab61-b067d19c1ce0" path="/var/lib/kubelet/pods/18497e75-1fc4-4324-ab61-b067d19c1ce0/volumes" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.360879 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.360936 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swg5c\" (UniqueName: \"kubernetes.io/projected/38ad16c2-10a9-4a20-a070-e1ed80f62684-kube-api-access-swg5c\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.360980 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.365011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.365082 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38ad16c2-10a9-4a20-a070-e1ed80f62684-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.389677 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swg5c\" (UniqueName: \"kubernetes.io/projected/38ad16c2-10a9-4a20-a070-e1ed80f62684-kube-api-access-swg5c\") pod \"nova-cell1-conductor-0\" (UID: \"38ad16c2-10a9-4a20-a070-e1ed80f62684\") " pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.442141 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.855687 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 20:04:09 crc kubenswrapper[4960]: W1201 20:04:09.946888 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38ad16c2_10a9_4a20_a070_e1ed80f62684.slice/crio-69bdb95e04ea236a9b33f2c3b832b980a1307356630ddea6525a5ae869e4bdc7 WatchSource:0}: Error finding container 69bdb95e04ea236a9b33f2c3b832b980a1307356630ddea6525a5ae869e4bdc7: Status 404 returned error can't find the container with id 69bdb95e04ea236a9b33f2c3b832b980a1307356630ddea6525a5ae869e4bdc7 Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.947996 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.975280 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-755hk\" (UniqueName: \"kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk\") pod \"7e7b2075-1317-4327-a8d9-4ebda5a68344\" (UID: \"7e7b2075-1317-4327-a8d9-4ebda5a68344\") " Dec 01 20:04:09 crc kubenswrapper[4960]: I1201 20:04:09.979214 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk" (OuterVolumeSpecName: "kube-api-access-755hk") pod "7e7b2075-1317-4327-a8d9-4ebda5a68344" (UID: "7e7b2075-1317-4327-a8d9-4ebda5a68344"). InnerVolumeSpecName "kube-api-access-755hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.056594 4960 generic.go:334] "Generic (PLEG): container finished" podID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerID="c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650" exitCode=2 Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.056893 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.056798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7e7b2075-1317-4327-a8d9-4ebda5a68344","Type":"ContainerDied","Data":"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650"} Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.057019 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7e7b2075-1317-4327-a8d9-4ebda5a68344","Type":"ContainerDied","Data":"7b63b8911a6cc3f1de2643e90ff71a126d94fa08c9ca751441271681db8c061a"} Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.057050 4960 scope.go:117] "RemoveContainer" containerID="c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.058441 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerName="nova-scheduler-scheduler" containerID="cri-o://8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" gracePeriod=30 Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.058620 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"38ad16c2-10a9-4a20-a070-e1ed80f62684","Type":"ContainerStarted","Data":"69bdb95e04ea236a9b33f2c3b832b980a1307356630ddea6525a5ae869e4bdc7"} Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.077484 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-755hk\" (UniqueName: \"kubernetes.io/projected/7e7b2075-1317-4327-a8d9-4ebda5a68344-kube-api-access-755hk\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.092437 4960 scope.go:117] "RemoveContainer" containerID="c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650" Dec 01 20:04:10 crc kubenswrapper[4960]: E1201 20:04:10.095209 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650\": container with ID starting with c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650 not found: ID does not exist" containerID="c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.095245 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650"} err="failed to get container status \"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650\": rpc error: code = NotFound desc = could not find container \"c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650\": container with ID starting with c121495a6316f2eed91347d6aa1aac28ab060c64f505a2da71c9a506fdd21650 not found: ID does not exist" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.103841 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.118665 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.135493 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:10 crc kubenswrapper[4960]: E1201 20:04:10.135981 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerName="kube-state-metrics" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.135999 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerName="kube-state-metrics" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.136230 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerName="kube-state-metrics" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.137006 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.140439 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.140467 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.145457 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.281373 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m284z\" (UniqueName: \"kubernetes.io/projected/e118a262-1b53-4d2a-adce-3802b626cff1-kube-api-access-m284z\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.281636 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.281705 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.281733 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.383573 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.383700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.383743 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.383767 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m284z\" (UniqueName: \"kubernetes.io/projected/e118a262-1b53-4d2a-adce-3802b626cff1-kube-api-access-m284z\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.388944 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.388961 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.389611 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e118a262-1b53-4d2a-adce-3802b626cff1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.415749 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m284z\" (UniqueName: \"kubernetes.io/projected/e118a262-1b53-4d2a-adce-3802b626cff1-kube-api-access-m284z\") pod \"kube-state-metrics-0\" (UID: \"e118a262-1b53-4d2a-adce-3802b626cff1\") " pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.454417 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 01 20:04:10 crc kubenswrapper[4960]: I1201 20:04:10.893031 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.072096 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e118a262-1b53-4d2a-adce-3802b626cff1","Type":"ContainerStarted","Data":"714aea31bb3f676ad2c4c5dd6c9fcc877cd80fa4e25d25c0018f6ea296a2f0b4"} Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.076330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"38ad16c2-10a9-4a20-a070-e1ed80f62684","Type":"ContainerStarted","Data":"844c4f7a1a034dc6ac3979dbfe0d2ed4e6a6cfa65103079b3f69b98f6e5fe071"} Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.076534 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.104280 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.10426142 podStartE2EDuration="2.10426142s" podCreationTimestamp="2025-12-01 20:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:11.098667015 +0000 UTC m=+1486.386158684" watchObservedRunningTime="2025-12-01 20:04:11.10426142 +0000 UTC m=+1486.391753089" Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.203330 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.204095 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-central-agent" containerID="cri-o://12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e" gracePeriod=30 Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.204135 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="proxy-httpd" containerID="cri-o://5bbf6281627287b92107ee9b051ae662a9d1aa9c0376dc47e7485aff04077593" gracePeriod=30 Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.204191 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="sg-core" containerID="cri-o://f86f8c0e3a4917ef1b4054f6d1150f6ea1c550e8f4787de54b2716f1649162cb" gracePeriod=30 Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.204246 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-notification-agent" containerID="cri-o://193dbfa7c1f3250c28521d277ce6db04be87cc53c5ec0411e2380b5c445da478" gracePeriod=30 Dec 01 20:04:11 crc kubenswrapper[4960]: I1201 20:04:11.341763 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" path="/var/lib/kubelet/pods/7e7b2075-1317-4327-a8d9-4ebda5a68344/volumes" Dec 01 20:04:11 crc kubenswrapper[4960]: E1201 20:04:11.749545 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c122f7_b0b2_4989_9984_3c3859eda76f.slice/crio-conmon-12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45c122f7_b0b2_4989_9984_3c3859eda76f.slice/crio-12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e.scope\": RecentStats: unable to find data in memory cache]" Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094568 4960 generic.go:334] "Generic (PLEG): container finished" podID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerID="5bbf6281627287b92107ee9b051ae662a9d1aa9c0376dc47e7485aff04077593" exitCode=0 Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094614 4960 generic.go:334] "Generic (PLEG): container finished" podID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerID="f86f8c0e3a4917ef1b4054f6d1150f6ea1c550e8f4787de54b2716f1649162cb" exitCode=2 Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094625 4960 generic.go:334] "Generic (PLEG): container finished" podID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerID="12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e" exitCode=0 Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerDied","Data":"5bbf6281627287b92107ee9b051ae662a9d1aa9c0376dc47e7485aff04077593"} Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094690 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerDied","Data":"f86f8c0e3a4917ef1b4054f6d1150f6ea1c550e8f4787de54b2716f1649162cb"} Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.094710 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerDied","Data":"12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e"} Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.100037 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e118a262-1b53-4d2a-adce-3802b626cff1","Type":"ContainerStarted","Data":"9784c9e8f222ef106e8b62bf2fa0323f9b1dd53ed54f8604ab68389fa5facc11"} Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.100144 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 01 20:04:12 crc kubenswrapper[4960]: I1201 20:04:12.123309 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.731803757 podStartE2EDuration="2.123271026s" podCreationTimestamp="2025-12-01 20:04:10 +0000 UTC" firstStartedPulling="2025-12-01 20:04:11.036987035 +0000 UTC m=+1486.324478714" lastFinishedPulling="2025-12-01 20:04:11.428454314 +0000 UTC m=+1486.715945983" observedRunningTime="2025-12-01 20:04:12.119520298 +0000 UTC m=+1487.407011967" watchObservedRunningTime="2025-12-01 20:04:12.123271026 +0000 UTC m=+1487.410762735" Dec 01 20:04:12 crc kubenswrapper[4960]: E1201 20:04:12.582895 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:12 crc kubenswrapper[4960]: E1201 20:04:12.585099 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:12 crc kubenswrapper[4960]: E1201 20:04:12.586617 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:12 crc kubenswrapper[4960]: E1201 20:04:12.586702 4960 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerName="nova-scheduler-scheduler" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.111826 4960 generic.go:334] "Generic (PLEG): container finished" podID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerID="193dbfa7c1f3250c28521d277ce6db04be87cc53c5ec0411e2380b5c445da478" exitCode=0 Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.112466 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerDied","Data":"193dbfa7c1f3250c28521d277ce6db04be87cc53c5ec0411e2380b5c445da478"} Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.495014 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.655972 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6djmm\" (UniqueName: \"kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656087 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656177 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656197 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656347 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656371 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.656400 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd\") pod \"45c122f7-b0b2-4989-9984-3c3859eda76f\" (UID: \"45c122f7-b0b2-4989-9984-3c3859eda76f\") " Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.657242 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.659443 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.663932 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm" (OuterVolumeSpecName: "kube-api-access-6djmm") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "kube-api-access-6djmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.684443 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts" (OuterVolumeSpecName: "scripts") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.713567 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.761080 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6djmm\" (UniqueName: \"kubernetes.io/projected/45c122f7-b0b2-4989-9984-3c3859eda76f-kube-api-access-6djmm\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.761130 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.761140 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.761149 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.761159 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/45c122f7-b0b2-4989-9984-3c3859eda76f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.765202 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.807461 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data" (OuterVolumeSpecName: "config-data") pod "45c122f7-b0b2-4989-9984-3c3859eda76f" (UID: "45c122f7-b0b2-4989-9984-3c3859eda76f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.862510 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.862553 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45c122f7-b0b2-4989-9984-3c3859eda76f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.870921 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.870989 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.871042 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.871896 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:04:13 crc kubenswrapper[4960]: I1201 20:04:13.871966 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923" gracePeriod=600 Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.127089 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.127094 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"45c122f7-b0b2-4989-9984-3c3859eda76f","Type":"ContainerDied","Data":"00854cdc42cf48f4c8b2ec625df0bae2764b9005327587a715e01263b0584233"} Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.127606 4960 scope.go:117] "RemoveContainer" containerID="5bbf6281627287b92107ee9b051ae662a9d1aa9c0376dc47e7485aff04077593" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.134745 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923" exitCode=0 Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.134808 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923"} Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.136810 4960 generic.go:334] "Generic (PLEG): container finished" podID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerID="8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" exitCode=0 Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.136835 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b5567cc-82b1-456a-ada3-2e6141808bfb","Type":"ContainerDied","Data":"8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d"} Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.160860 4960 scope.go:117] "RemoveContainer" containerID="f86f8c0e3a4917ef1b4054f6d1150f6ea1c550e8f4787de54b2716f1649162cb" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.205255 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.219216 4960 scope.go:117] "RemoveContainer" containerID="193dbfa7c1f3250c28521d277ce6db04be87cc53c5ec0411e2380b5c445da478" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.232644 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.245216 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:14 crc kubenswrapper[4960]: E1201 20:04:14.245969 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-central-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.245983 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-central-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: E1201 20:04:14.245999 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="proxy-httpd" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246005 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="proxy-httpd" Dec 01 20:04:14 crc kubenswrapper[4960]: E1201 20:04:14.246027 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-notification-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246033 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-notification-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: E1201 20:04:14.246051 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="sg-core" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246056 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="sg-core" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246269 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="proxy-httpd" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246288 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-central-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246304 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="sg-core" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.246320 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" containerName="ceilometer-notification-agent" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.248988 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.251260 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.255478 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.255733 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.259432 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.274303 4960 scope.go:117] "RemoveContainer" containerID="12ddd713981f522ef0e29c860a4fd138447ed522adb68b2f9a30f19a71f2548e" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.303051 4960 scope.go:117] "RemoveContainer" containerID="ddf6a869331c4dc13e30fdb64aa28bddb51cd09bb6ecb075855ace122b423e07" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375563 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375606 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375626 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bxp\" (UniqueName: \"kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375656 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375697 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375867 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.375904 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.465398 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.477969 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478098 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478197 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478238 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478281 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478301 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bxp\" (UniqueName: \"kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478352 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478582 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.478607 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.487021 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.487023 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.487218 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.492891 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.494613 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.494770 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="7e7b2075-1317-4327-a8d9-4ebda5a68344" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": dial tcp 10.217.0.112:8081: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.501947 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bxp\" (UniqueName: \"kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp\") pod \"ceilometer-0\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.575708 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.579634 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle\") pod \"4b5567cc-82b1-456a-ada3-2e6141808bfb\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.579780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdk8m\" (UniqueName: \"kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m\") pod \"4b5567cc-82b1-456a-ada3-2e6141808bfb\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.579847 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data\") pod \"4b5567cc-82b1-456a-ada3-2e6141808bfb\" (UID: \"4b5567cc-82b1-456a-ada3-2e6141808bfb\") " Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.590423 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m" (OuterVolumeSpecName: "kube-api-access-hdk8m") pod "4b5567cc-82b1-456a-ada3-2e6141808bfb" (UID: "4b5567cc-82b1-456a-ada3-2e6141808bfb"). InnerVolumeSpecName "kube-api-access-hdk8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.624372 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data" (OuterVolumeSpecName: "config-data") pod "4b5567cc-82b1-456a-ada3-2e6141808bfb" (UID: "4b5567cc-82b1-456a-ada3-2e6141808bfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.625143 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b5567cc-82b1-456a-ada3-2e6141808bfb" (UID: "4b5567cc-82b1-456a-ada3-2e6141808bfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.681712 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.681742 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdk8m\" (UniqueName: \"kubernetes.io/projected/4b5567cc-82b1-456a-ada3-2e6141808bfb-kube-api-access-hdk8m\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:14 crc kubenswrapper[4960]: I1201 20:04:14.681756 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b5567cc-82b1-456a-ada3-2e6141808bfb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.170428 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.175651 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01"} Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.184274 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.186360 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b5567cc-82b1-456a-ada3-2e6141808bfb","Type":"ContainerDied","Data":"e0f78c7c2e3576959b46393eeacab95f25cf36b51f47533ba75131064a110dc3"} Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.186432 4960 scope.go:117] "RemoveContainer" containerID="8b5ac4100b46783e36886221b5c4211d32c01a947ca8a302e8c2c80a9563ec8d" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.191189 4960 generic.go:334] "Generic (PLEG): container finished" podID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerID="a0fd097755e2fabb955364cdcc48f93c04efbe40c5789108e55ca03ee0edc2cd" exitCode=0 Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.191265 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerDied","Data":"a0fd097755e2fabb955364cdcc48f93c04efbe40c5789108e55ca03ee0edc2cd"} Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.238153 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.255808 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.264038 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:15 crc kubenswrapper[4960]: E1201 20:04:15.264613 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerName="nova-scheduler-scheduler" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.264627 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerName="nova-scheduler-scheduler" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.264838 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" containerName="nova-scheduler-scheduler" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.265645 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.268808 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.272345 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.337558 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45c122f7-b0b2-4989-9984-3c3859eda76f" path="/var/lib/kubelet/pods/45c122f7-b0b2-4989-9984-3c3859eda76f/volumes" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.338366 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b5567cc-82b1-456a-ada3-2e6141808bfb" path="/var/lib/kubelet/pods/4b5567cc-82b1-456a-ada3-2e6141808bfb/volumes" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.373716 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.394490 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg6ls\" (UniqueName: \"kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.394714 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.394750 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.495893 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data\") pod \"9c393fd6-f1d0-4c53-8558-e909279da2de\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.496298 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle\") pod \"9c393fd6-f1d0-4c53-8558-e909279da2de\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.496407 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs\") pod \"9c393fd6-f1d0-4c53-8558-e909279da2de\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.496516 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjt5m\" (UniqueName: \"kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m\") pod \"9c393fd6-f1d0-4c53-8558-e909279da2de\" (UID: \"9c393fd6-f1d0-4c53-8558-e909279da2de\") " Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.496892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.496921 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.497011 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg6ls\" (UniqueName: \"kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.497916 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs" (OuterVolumeSpecName: "logs") pod "9c393fd6-f1d0-4c53-8558-e909279da2de" (UID: "9c393fd6-f1d0-4c53-8558-e909279da2de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.502651 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.503153 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m" (OuterVolumeSpecName: "kube-api-access-zjt5m") pod "9c393fd6-f1d0-4c53-8558-e909279da2de" (UID: "9c393fd6-f1d0-4c53-8558-e909279da2de"). InnerVolumeSpecName "kube-api-access-zjt5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.504704 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.516082 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg6ls\" (UniqueName: \"kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls\") pod \"nova-scheduler-0\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " pod="openstack/nova-scheduler-0" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.526063 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data" (OuterVolumeSpecName: "config-data") pod "9c393fd6-f1d0-4c53-8558-e909279da2de" (UID: "9c393fd6-f1d0-4c53-8558-e909279da2de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.526676 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c393fd6-f1d0-4c53-8558-e909279da2de" (UID: "9c393fd6-f1d0-4c53-8558-e909279da2de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.599547 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.599585 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c393fd6-f1d0-4c53-8558-e909279da2de-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.599594 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjt5m\" (UniqueName: \"kubernetes.io/projected/9c393fd6-f1d0-4c53-8558-e909279da2de-kube-api-access-zjt5m\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.599606 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c393fd6-f1d0-4c53-8558-e909279da2de-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:15 crc kubenswrapper[4960]: I1201 20:04:15.668658 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.155886 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.209423 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9c393fd6-f1d0-4c53-8558-e909279da2de","Type":"ContainerDied","Data":"7eb35947787275e964d03f31d4e3fa3b9cda0e6e28d165860c3f0f9132f40ca7"} Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.209484 4960 scope.go:117] "RemoveContainer" containerID="a0fd097755e2fabb955364cdcc48f93c04efbe40c5789108e55ca03ee0edc2cd" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.209672 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.213737 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerStarted","Data":"329b31d92e1e4cf6f0f879d1325dd461cd38b2d9037561808fff26c3698c5308"} Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.216226 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"20d77553-9219-4b2c-b0fc-5c642600a604","Type":"ContainerStarted","Data":"b82bd8250716f434241d163a5a316707a728867ebc25f070e6558ae64d097fee"} Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.300844 4960 scope.go:117] "RemoveContainer" containerID="f38da714308b5ddcb0bddb39b8781115a23ffdbafdf892341449fd1889d3410e" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.346521 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.361295 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.371575 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:16 crc kubenswrapper[4960]: E1201 20:04:16.372326 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-api" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.372346 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-api" Dec 01 20:04:16 crc kubenswrapper[4960]: E1201 20:04:16.372415 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-log" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.372424 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-log" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.372781 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-log" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.372811 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" containerName="nova-api-api" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.374599 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.379486 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.385527 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.522748 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.522931 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.522986 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5rp5\" (UniqueName: \"kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.523190 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.626063 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.626528 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.626701 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.626753 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5rp5\" (UniqueName: \"kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.626913 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.632585 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.632830 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.645822 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5rp5\" (UniqueName: \"kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5\") pod \"nova-api-0\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " pod="openstack/nova-api-0" Dec 01 20:04:16 crc kubenswrapper[4960]: I1201 20:04:16.698882 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:17 crc kubenswrapper[4960]: W1201 20:04:17.156618 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ad45e60_054e_4654_ae3b_7cf69536160f.slice/crio-d30d629b640cbaed1059c1c306293de711eabb465adc423d528c1844d782a6aa WatchSource:0}: Error finding container d30d629b640cbaed1059c1c306293de711eabb465adc423d528c1844d782a6aa: Status 404 returned error can't find the container with id d30d629b640cbaed1059c1c306293de711eabb465adc423d528c1844d782a6aa Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.167205 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.228813 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerStarted","Data":"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750"} Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.230905 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"20d77553-9219-4b2c-b0fc-5c642600a604","Type":"ContainerStarted","Data":"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599"} Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.232182 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerStarted","Data":"d30d629b640cbaed1059c1c306293de711eabb465adc423d528c1844d782a6aa"} Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.253286 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.253268405 podStartE2EDuration="2.253268405s" podCreationTimestamp="2025-12-01 20:04:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:17.251316261 +0000 UTC m=+1492.538807950" watchObservedRunningTime="2025-12-01 20:04:17.253268405 +0000 UTC m=+1492.540760064" Dec 01 20:04:17 crc kubenswrapper[4960]: I1201 20:04:17.340055 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c393fd6-f1d0-4c53-8558-e909279da2de" path="/var/lib/kubelet/pods/9c393fd6-f1d0-4c53-8558-e909279da2de/volumes" Dec 01 20:04:18 crc kubenswrapper[4960]: I1201 20:04:18.245062 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerStarted","Data":"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5"} Dec 01 20:04:18 crc kubenswrapper[4960]: I1201 20:04:18.248800 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerStarted","Data":"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe"} Dec 01 20:04:18 crc kubenswrapper[4960]: I1201 20:04:18.248839 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerStarted","Data":"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2"} Dec 01 20:04:18 crc kubenswrapper[4960]: I1201 20:04:18.272351 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.272327223 podStartE2EDuration="2.272327223s" podCreationTimestamp="2025-12-01 20:04:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:18.268367254 +0000 UTC m=+1493.555858923" watchObservedRunningTime="2025-12-01 20:04:18.272327223 +0000 UTC m=+1493.559818892" Dec 01 20:04:19 crc kubenswrapper[4960]: I1201 20:04:19.265354 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerStarted","Data":"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8"} Dec 01 20:04:19 crc kubenswrapper[4960]: I1201 20:04:19.481480 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 01 20:04:20 crc kubenswrapper[4960]: I1201 20:04:20.278790 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerStarted","Data":"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02"} Dec 01 20:04:20 crc kubenswrapper[4960]: I1201 20:04:20.279443 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:04:20 crc kubenswrapper[4960]: I1201 20:04:20.312817 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.504123183 podStartE2EDuration="6.312789045s" podCreationTimestamp="2025-12-01 20:04:14 +0000 UTC" firstStartedPulling="2025-12-01 20:04:15.178507157 +0000 UTC m=+1490.465998816" lastFinishedPulling="2025-12-01 20:04:19.987172999 +0000 UTC m=+1495.274664678" observedRunningTime="2025-12-01 20:04:20.308685721 +0000 UTC m=+1495.596177430" watchObservedRunningTime="2025-12-01 20:04:20.312789045 +0000 UTC m=+1495.600280734" Dec 01 20:04:20 crc kubenswrapper[4960]: I1201 20:04:20.463707 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 01 20:04:20 crc kubenswrapper[4960]: I1201 20:04:20.669433 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 20:04:25 crc kubenswrapper[4960]: I1201 20:04:25.669243 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 20:04:25 crc kubenswrapper[4960]: I1201 20:04:25.725605 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 20:04:26 crc kubenswrapper[4960]: I1201 20:04:26.374759 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 20:04:26 crc kubenswrapper[4960]: I1201 20:04:26.700396 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:04:26 crc kubenswrapper[4960]: I1201 20:04:26.700459 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:04:27 crc kubenswrapper[4960]: I1201 20:04:27.782444 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.219:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 20:04:27 crc kubenswrapper[4960]: I1201 20:04:27.782834 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.219:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 01 20:04:33 crc kubenswrapper[4960]: I1201 20:04:33.422062 4960 generic.go:334] "Generic (PLEG): container finished" podID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" containerID="0400687f51e97a3e36fbed612e1c6cc0fd75180255637a22ce2f9004419fdc9f" exitCode=137 Dec 01 20:04:33 crc kubenswrapper[4960]: I1201 20:04:33.422174 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5","Type":"ContainerDied","Data":"0400687f51e97a3e36fbed612e1c6cc0fd75180255637a22ce2f9004419fdc9f"} Dec 01 20:04:33 crc kubenswrapper[4960]: I1201 20:04:33.423898 4960 generic.go:334] "Generic (PLEG): container finished" podID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerID="516807a73e40f0edb3f2e5e1c2fed6d126ddb1a2e51b77f1f4fe8bc9a1d0ae33" exitCode=137 Dec 01 20:04:33 crc kubenswrapper[4960]: I1201 20:04:33.423922 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerDied","Data":"516807a73e40f0edb3f2e5e1c2fed6d126ddb1a2e51b77f1f4fe8bc9a1d0ae33"} Dec 01 20:04:33 crc kubenswrapper[4960]: I1201 20:04:33.994887 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.000285 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142292 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle\") pod \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142352 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6qxn\" (UniqueName: \"kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn\") pod \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142432 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data\") pod \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142510 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b776b\" (UniqueName: \"kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b\") pod \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142536 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle\") pod \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142601 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data\") pod \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\" (UID: \"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.142637 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs\") pod \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\" (UID: \"fc6382cf-fc5a-4601-a40d-7b34426cb73e\") " Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.143641 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs" (OuterVolumeSpecName: "logs") pod "fc6382cf-fc5a-4601-a40d-7b34426cb73e" (UID: "fc6382cf-fc5a-4601-a40d-7b34426cb73e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.148145 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn" (OuterVolumeSpecName: "kube-api-access-c6qxn") pod "fc6382cf-fc5a-4601-a40d-7b34426cb73e" (UID: "fc6382cf-fc5a-4601-a40d-7b34426cb73e"). InnerVolumeSpecName "kube-api-access-c6qxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.148335 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b" (OuterVolumeSpecName: "kube-api-access-b776b") pod "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" (UID: "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5"). InnerVolumeSpecName "kube-api-access-b776b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.171715 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc6382cf-fc5a-4601-a40d-7b34426cb73e" (UID: "fc6382cf-fc5a-4601-a40d-7b34426cb73e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.173325 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data" (OuterVolumeSpecName: "config-data") pod "fc6382cf-fc5a-4601-a40d-7b34426cb73e" (UID: "fc6382cf-fc5a-4601-a40d-7b34426cb73e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.175279 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" (UID: "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.197357 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data" (OuterVolumeSpecName: "config-data") pod "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" (UID: "fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251590 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251654 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc6382cf-fc5a-4601-a40d-7b34426cb73e-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251666 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251698 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6qxn\" (UniqueName: \"kubernetes.io/projected/fc6382cf-fc5a-4601-a40d-7b34426cb73e-kube-api-access-c6qxn\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251709 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc6382cf-fc5a-4601-a40d-7b34426cb73e-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251721 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b776b\" (UniqueName: \"kubernetes.io/projected/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-kube-api-access-b776b\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.251730 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.437107 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fc6382cf-fc5a-4601-a40d-7b34426cb73e","Type":"ContainerDied","Data":"1088a7b00e6db6b3d8acb74246a2fbcdff84cb491142e09b7d77df1f38344cb3"} Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.437167 4960 scope.go:117] "RemoveContainer" containerID="516807a73e40f0edb3f2e5e1c2fed6d126ddb1a2e51b77f1f4fe8bc9a1d0ae33" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.437230 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.439019 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5","Type":"ContainerDied","Data":"f08334c2f87c54e65965a87163130db90706f890e0bc8f8d64176eef9144221d"} Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.439050 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.478866 4960 scope.go:117] "RemoveContainer" containerID="a912d7be53949fe393f27f58542ae90ff94092398786345c76715fac25517052" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.545017 4960 scope.go:117] "RemoveContainer" containerID="0400687f51e97a3e36fbed612e1c6cc0fd75180255637a22ce2f9004419fdc9f" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.555245 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.566783 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.578394 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.593095 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.604729 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: E1201 20:04:34.605412 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-metadata" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605437 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-metadata" Dec 01 20:04:34 crc kubenswrapper[4960]: E1201 20:04:34.605476 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-log" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605486 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-log" Dec 01 20:04:34 crc kubenswrapper[4960]: E1201 20:04:34.605513 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605555 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605785 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-metadata" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605836 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" containerName="nova-metadata-log" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.605850 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" containerName="nova-cell1-novncproxy-novncproxy" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.607701 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.610762 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.611022 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.611229 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.623808 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.633372 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.635546 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.639027 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.639092 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.642681 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.761705 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z972m\" (UniqueName: \"kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762084 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762238 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762281 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762374 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762410 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762444 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rgkx\" (UniqueName: \"kubernetes.io/projected/35daed02-01d5-4757-aa71-9c38e1397d19-kube-api-access-5rgkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762471 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.762501 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864215 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864327 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864355 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864383 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864402 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864419 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864438 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rgkx\" (UniqueName: \"kubernetes.io/projected/35daed02-01d5-4757-aa71-9c38e1397d19-kube-api-access-5rgkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864471 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864496 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864810 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z972m\" (UniqueName: \"kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.864936 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.871249 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.871377 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.871473 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.872425 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.873541 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.874643 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/35daed02-01d5-4757-aa71-9c38e1397d19-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.874903 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.880860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rgkx\" (UniqueName: \"kubernetes.io/projected/35daed02-01d5-4757-aa71-9c38e1397d19-kube-api-access-5rgkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"35daed02-01d5-4757-aa71-9c38e1397d19\") " pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.882282 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z972m\" (UniqueName: \"kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m\") pod \"nova-metadata-0\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " pod="openstack/nova-metadata-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.942423 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:34 crc kubenswrapper[4960]: I1201 20:04:34.957240 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:04:35 crc kubenswrapper[4960]: I1201 20:04:35.341656 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc6382cf-fc5a-4601-a40d-7b34426cb73e" path="/var/lib/kubelet/pods/fc6382cf-fc5a-4601-a40d-7b34426cb73e/volumes" Dec 01 20:04:35 crc kubenswrapper[4960]: I1201 20:04:35.343042 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5" path="/var/lib/kubelet/pods/fe1e685d-8f3b-48ff-8d5a-776ba4fb9fa5/volumes" Dec 01 20:04:35 crc kubenswrapper[4960]: I1201 20:04:35.429240 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:35 crc kubenswrapper[4960]: W1201 20:04:35.434094 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod319bb370_d760_40b2_89c2_e52d1ff260d5.slice/crio-d097f90f454e99aa850ed2e7fc7b03f551d8bfb35673eeb01a0f1317dd08c240 WatchSource:0}: Error finding container d097f90f454e99aa850ed2e7fc7b03f551d8bfb35673eeb01a0f1317dd08c240: Status 404 returned error can't find the container with id d097f90f454e99aa850ed2e7fc7b03f551d8bfb35673eeb01a0f1317dd08c240 Dec 01 20:04:35 crc kubenswrapper[4960]: I1201 20:04:35.458317 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 01 20:04:35 crc kubenswrapper[4960]: I1201 20:04:35.468955 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerStarted","Data":"d097f90f454e99aa850ed2e7fc7b03f551d8bfb35673eeb01a0f1317dd08c240"} Dec 01 20:04:35 crc kubenswrapper[4960]: W1201 20:04:35.473794 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35daed02_01d5_4757_aa71_9c38e1397d19.slice/crio-4c61036f63c1ebe746573e88eaf351d1f1506a9e87b458126e5bcf495a8c6f5b WatchSource:0}: Error finding container 4c61036f63c1ebe746573e88eaf351d1f1506a9e87b458126e5bcf495a8c6f5b: Status 404 returned error can't find the container with id 4c61036f63c1ebe746573e88eaf351d1f1506a9e87b458126e5bcf495a8c6f5b Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.481238 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"35daed02-01d5-4757-aa71-9c38e1397d19","Type":"ContainerStarted","Data":"8fccbbc4f2f58abc8ed5de29f7255bf4f7a1ef1ddedb59657a8e9fd648065ab8"} Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.481530 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"35daed02-01d5-4757-aa71-9c38e1397d19","Type":"ContainerStarted","Data":"4c61036f63c1ebe746573e88eaf351d1f1506a9e87b458126e5bcf495a8c6f5b"} Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.484377 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerStarted","Data":"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e"} Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.484437 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerStarted","Data":"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c"} Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.513511 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.513490013 podStartE2EDuration="2.513490013s" podCreationTimestamp="2025-12-01 20:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:36.503582261 +0000 UTC m=+1511.791073930" watchObservedRunningTime="2025-12-01 20:04:36.513490013 +0000 UTC m=+1511.800981692" Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.532585 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.5325626530000003 podStartE2EDuration="2.532562653s" podCreationTimestamp="2025-12-01 20:04:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:36.5214064 +0000 UTC m=+1511.808898069" watchObservedRunningTime="2025-12-01 20:04:36.532562653 +0000 UTC m=+1511.820054342" Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.705818 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.707330 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.707408 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 20:04:36 crc kubenswrapper[4960]: I1201 20:04:36.711549 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.494073 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.500449 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.679060 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.681504 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.707716 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829540 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829623 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829716 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829769 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829880 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v656\" (UniqueName: \"kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.829930 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931284 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931405 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931440 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931469 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931503 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.931599 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v656\" (UniqueName: \"kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.932998 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.933006 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.933047 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.933164 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.933565 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:37 crc kubenswrapper[4960]: I1201 20:04:37.950165 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v656\" (UniqueName: \"kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656\") pod \"dnsmasq-dns-5fd9b586ff-rh624\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:38 crc kubenswrapper[4960]: I1201 20:04:38.015763 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:38 crc kubenswrapper[4960]: I1201 20:04:38.510207 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:04:38 crc kubenswrapper[4960]: W1201 20:04:38.511471 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17246315_db1a_4a54_ae06_2060ce778508.slice/crio-c84aff5d2a2b5b510dd4508809b77b6b6bb4679335bda84cd4b4c28a0c21a63d WatchSource:0}: Error finding container c84aff5d2a2b5b510dd4508809b77b6b6bb4679335bda84cd4b4c28a0c21a63d: Status 404 returned error can't find the container with id c84aff5d2a2b5b510dd4508809b77b6b6bb4679335bda84cd4b4c28a0c21a63d Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.513145 4960 generic.go:334] "Generic (PLEG): container finished" podID="17246315-db1a-4a54-ae06-2060ce778508" containerID="1ed6d867f843e39a4d3a63e8ee631e90d76949c4ce088870071d0acbe9bf9362" exitCode=0 Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.513338 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" event={"ID":"17246315-db1a-4a54-ae06-2060ce778508","Type":"ContainerDied","Data":"1ed6d867f843e39a4d3a63e8ee631e90d76949c4ce088870071d0acbe9bf9362"} Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.514451 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" event={"ID":"17246315-db1a-4a54-ae06-2060ce778508","Type":"ContainerStarted","Data":"c84aff5d2a2b5b510dd4508809b77b6b6bb4679335bda84cd4b4c28a0c21a63d"} Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.842922 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.843283 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-central-agent" containerID="cri-o://18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750" gracePeriod=30 Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.843445 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="proxy-httpd" containerID="cri-o://c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02" gracePeriod=30 Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.843510 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="sg-core" containerID="cri-o://f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8" gracePeriod=30 Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.843568 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-notification-agent" containerID="cri-o://85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5" gracePeriod=30 Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.857960 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.217:3000/\": EOF" Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.943510 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.958491 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:04:39 crc kubenswrapper[4960]: I1201 20:04:39.958546 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.160957 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.529343 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" event={"ID":"17246315-db1a-4a54-ae06-2060ce778508","Type":"ContainerStarted","Data":"54d0f458c30b5f5153d1e15cbf8ee70dee2e36af91ccd718c542f3d7cd0d8b79"} Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.532125 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.537728 4960 generic.go:334] "Generic (PLEG): container finished" podID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerID="c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02" exitCode=0 Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.537921 4960 generic.go:334] "Generic (PLEG): container finished" podID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerID="f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8" exitCode=2 Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.537985 4960 generic.go:334] "Generic (PLEG): container finished" podID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerID="18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750" exitCode=0 Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.538070 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerDied","Data":"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02"} Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.538167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerDied","Data":"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8"} Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.538187 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerDied","Data":"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750"} Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.538378 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-log" containerID="cri-o://bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2" gracePeriod=30 Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.538544 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-api" containerID="cri-o://05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe" gracePeriod=30 Dec 01 20:04:40 crc kubenswrapper[4960]: I1201 20:04:40.562760 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" podStartSLOduration=3.562739349 podStartE2EDuration="3.562739349s" podCreationTimestamp="2025-12-01 20:04:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:40.547796152 +0000 UTC m=+1515.835287831" watchObservedRunningTime="2025-12-01 20:04:40.562739349 +0000 UTC m=+1515.850231008" Dec 01 20:04:41 crc kubenswrapper[4960]: I1201 20:04:41.550945 4960 generic.go:334] "Generic (PLEG): container finished" podID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerID="bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2" exitCode=143 Dec 01 20:04:41 crc kubenswrapper[4960]: I1201 20:04:41.551045 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerDied","Data":"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2"} Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.392364 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.463891 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.464321 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="sg-core" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464340 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="sg-core" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.464364 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="proxy-httpd" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464372 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="proxy-httpd" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.464391 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-central-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464397 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-central-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.464422 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-notification-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464429 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-notification-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464603 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-central-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464624 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="sg-core" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464638 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="ceilometer-notification-agent" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.464647 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerName="proxy-httpd" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.466151 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.476248 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.572802 4960 generic.go:334] "Generic (PLEG): container finished" podID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" containerID="85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5" exitCode=0 Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.572886 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.572913 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerDied","Data":"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5"} Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.573358 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fbf4ea6e-e9b8-451a-8b15-78fddd95c462","Type":"ContainerDied","Data":"329b31d92e1e4cf6f0f879d1325dd461cd38b2d9037561808fff26c3698c5308"} Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.573381 4960 scope.go:117] "RemoveContainer" containerID="c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588068 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588195 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588309 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588347 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588383 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588770 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.588800 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bxp\" (UniqueName: \"kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp\") pod \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\" (UID: \"fbf4ea6e-e9b8-451a-8b15-78fddd95c462\") " Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.589149 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.589178 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.589249 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szh7x\" (UniqueName: \"kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.590733 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.590907 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.603999 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp" (OuterVolumeSpecName: "kube-api-access-97bxp") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "kube-api-access-97bxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.608417 4960 scope.go:117] "RemoveContainer" containerID="f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.625495 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts" (OuterVolumeSpecName: "scripts") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.643164 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.690028 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.691909 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szh7x\" (UniqueName: \"kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692373 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692502 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692842 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bxp\" (UniqueName: \"kubernetes.io/projected/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-kube-api-access-97bxp\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692867 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692981 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.693000 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.693012 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.693023 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.692836 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.693436 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.712128 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.712504 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szh7x\" (UniqueName: \"kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x\") pod \"certified-operators-dq66q\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.779641 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data" (OuterVolumeSpecName: "config-data") pod "fbf4ea6e-e9b8-451a-8b15-78fddd95c462" (UID: "fbf4ea6e-e9b8-451a-8b15-78fddd95c462"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.784052 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.795001 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.795027 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf4ea6e-e9b8-451a-8b15-78fddd95c462-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.846693 4960 scope.go:117] "RemoveContainer" containerID="85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.885673 4960 scope.go:117] "RemoveContainer" containerID="18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.916631 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.918981 4960 scope.go:117] "RemoveContainer" containerID="c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.919402 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02\": container with ID starting with c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02 not found: ID does not exist" containerID="c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.919430 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02"} err="failed to get container status \"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02\": rpc error: code = NotFound desc = could not find container \"c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02\": container with ID starting with c2762fe0fb25ea6c746fd2e89fe5ac12990538009a0a72072f833a5feeb5bf02 not found: ID does not exist" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.919451 4960 scope.go:117] "RemoveContainer" containerID="f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.920068 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8\": container with ID starting with f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8 not found: ID does not exist" containerID="f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.925549 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8"} err="failed to get container status \"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8\": rpc error: code = NotFound desc = could not find container \"f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8\": container with ID starting with f36059278281a403ac480355e7af9d12acdfe316dcfbd85ebd4e8dec54611ec8 not found: ID does not exist" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.925607 4960 scope.go:117] "RemoveContainer" containerID="85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.935821 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5\": container with ID starting with 85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5 not found: ID does not exist" containerID="85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.935870 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5"} err="failed to get container status \"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5\": rpc error: code = NotFound desc = could not find container \"85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5\": container with ID starting with 85780619f6727573940084fcc89373a282e38bd0af60d1ae7465798845a82fa5 not found: ID does not exist" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.935905 4960 scope.go:117] "RemoveContainer" containerID="18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750" Dec 01 20:04:43 crc kubenswrapper[4960]: E1201 20:04:43.943248 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750\": container with ID starting with 18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750 not found: ID does not exist" containerID="18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.943545 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750"} err="failed to get container status \"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750\": rpc error: code = NotFound desc = could not find container \"18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750\": container with ID starting with 18fda7e391aee0c8573709f79dfaa9cd0f0976baee2d2fe192bfe23aa059b750 not found: ID does not exist" Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.954751 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:43 crc kubenswrapper[4960]: I1201 20:04:43.990436 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:43.994802 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.005749 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.005889 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.005929 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.005964 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.006008 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.006066 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmnjk\" (UniqueName: \"kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.006090 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.006171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.013757 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.014015 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.014161 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.048927 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116615 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116672 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmnjk\" (UniqueName: \"kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116691 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116720 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116779 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116844 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116870 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.116891 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.118798 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.119080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.141731 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.154807 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.162943 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.175017 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.178974 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.179220 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmnjk\" (UniqueName: \"kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk\") pod \"ceilometer-0\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.364375 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.396844 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.428314 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data\") pod \"7ad45e60-054e-4654-ae3b-7cf69536160f\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.428379 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs\") pod \"7ad45e60-054e-4654-ae3b-7cf69536160f\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.428481 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle\") pod \"7ad45e60-054e-4654-ae3b-7cf69536160f\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.428502 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5rp5\" (UniqueName: \"kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5\") pod \"7ad45e60-054e-4654-ae3b-7cf69536160f\" (UID: \"7ad45e60-054e-4654-ae3b-7cf69536160f\") " Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.429759 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs" (OuterVolumeSpecName: "logs") pod "7ad45e60-054e-4654-ae3b-7cf69536160f" (UID: "7ad45e60-054e-4654-ae3b-7cf69536160f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.446282 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5" (OuterVolumeSpecName: "kube-api-access-g5rp5") pod "7ad45e60-054e-4654-ae3b-7cf69536160f" (UID: "7ad45e60-054e-4654-ae3b-7cf69536160f"). InnerVolumeSpecName "kube-api-access-g5rp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.458518 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.466324 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ad45e60-054e-4654-ae3b-7cf69536160f" (UID: "7ad45e60-054e-4654-ae3b-7cf69536160f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.474576 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data" (OuterVolumeSpecName: "config-data") pod "7ad45e60-054e-4654-ae3b-7cf69536160f" (UID: "7ad45e60-054e-4654-ae3b-7cf69536160f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.532435 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.533069 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad45e60-054e-4654-ae3b-7cf69536160f-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.533081 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad45e60-054e-4654-ae3b-7cf69536160f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:44 crc kubenswrapper[4960]: I1201 20:04:44.533093 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5rp5\" (UniqueName: \"kubernetes.io/projected/7ad45e60-054e-4654-ae3b-7cf69536160f-kube-api-access-g5rp5\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.592353 4960 generic.go:334] "Generic (PLEG): container finished" podID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerID="05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe" exitCode=0 Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.592414 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerDied","Data":"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe"} Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.592430 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.592453 4960 scope.go:117] "RemoveContainer" containerID="05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.592440 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7ad45e60-054e-4654-ae3b-7cf69536160f","Type":"ContainerDied","Data":"d30d629b640cbaed1059c1c306293de711eabb465adc423d528c1844d782a6aa"} Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.594277 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerStarted","Data":"922e273e621f88ee3df9b7acb7f5ba1ae369595d1a36250c380b4d20684f9e57"} Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.615269 4960 scope.go:117] "RemoveContainer" containerID="bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.656464 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.680139 4960 scope.go:117] "RemoveContainer" containerID="05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe" Dec 01 20:04:45 crc kubenswrapper[4960]: E1201 20:04:44.680655 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe\": container with ID starting with 05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe not found: ID does not exist" containerID="05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.680679 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe"} err="failed to get container status \"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe\": rpc error: code = NotFound desc = could not find container \"05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe\": container with ID starting with 05575fbfc2d8fab55bbd9f8154a21b3dfd2050c9e2e129b7ab9ab02c6ecd30fe not found: ID does not exist" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.680701 4960 scope.go:117] "RemoveContainer" containerID="bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2" Dec 01 20:04:45 crc kubenswrapper[4960]: E1201 20:04:44.681077 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2\": container with ID starting with bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2 not found: ID does not exist" containerID="bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.681249 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2"} err="failed to get container status \"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2\": rpc error: code = NotFound desc = could not find container \"bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2\": container with ID starting with bbdec3264b01e9c8ddd178c7eab0f987ee2abd7e052c74564fe05d287d079ed2 not found: ID does not exist" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.686485 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.696454 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: E1201 20:04:44.696853 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-api" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.696865 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-api" Dec 01 20:04:45 crc kubenswrapper[4960]: E1201 20:04:44.696908 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-log" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.696916 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-log" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.697096 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-log" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.697145 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" containerName="nova-api-api" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.698622 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.700909 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.704058 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.704447 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.709157 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840607 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsj9c\" (UniqueName: \"kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840735 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840777 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840807 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840830 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.840865 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.841713 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.844412 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.856695 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.942745 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.942927 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.943199 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsj9c\" (UniqueName: \"kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.943360 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.943413 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.943452 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.943477 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.944019 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.950819 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.950882 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.950988 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.951818 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.957940 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.959024 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:44.963264 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsj9c\" (UniqueName: \"kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c\") pod \"nova-api-0\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.007893 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.045832 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.046926 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4jdb\" (UniqueName: \"kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.047059 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.069641 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.149357 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.149494 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.149555 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4jdb\" (UniqueName: \"kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.149898 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.149914 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.172371 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4jdb\" (UniqueName: \"kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb\") pod \"redhat-operators-bdbr2\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.173579 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.340750 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad45e60-054e-4654-ae3b-7cf69536160f" path="/var/lib/kubelet/pods/7ad45e60-054e-4654-ae3b-7cf69536160f/volumes" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.341641 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf4ea6e-e9b8-451a-8b15-78fddd95c462" path="/var/lib/kubelet/pods/fbf4ea6e-e9b8-451a-8b15-78fddd95c462/volumes" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.523703 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.606914 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerStarted","Data":"ea44aaee72053234872a11ce12e4e1605850ae5de7541d74ccb63e387e78dfb1"} Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.608846 4960 generic.go:334] "Generic (PLEG): container finished" podID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerID="4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7" exitCode=0 Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.608917 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerDied","Data":"4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7"} Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.637821 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.837260 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.897263 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.965741 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-tvqfm"] Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.967227 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.968395 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.970674 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.975147 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.977537 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:04:45 crc kubenswrapper[4960]: I1201 20:04:45.990804 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tvqfm"] Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.092220 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.092412 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.092441 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.092474 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9r6n\" (UniqueName: \"kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.195039 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.195602 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.195688 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9r6n\" (UniqueName: \"kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.195970 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.213567 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.220306 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.224838 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.226724 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9r6n\" (UniqueName: \"kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n\") pod \"nova-cell1-cell-mapping-tvqfm\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.417307 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.646348 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerStarted","Data":"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.671869 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerStarted","Data":"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.686732 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerStarted","Data":"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.686777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerStarted","Data":"14723fcd585a1fa936b8d9bd46aadaf9af9085601a65ff2f8e3847aa1c9c9c0c"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.699287 4960 generic.go:334] "Generic (PLEG): container finished" podID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerID="ff28f8e07e09ab9cd1123e1820ca7f55f0fdb1ecc1dd405fa046a7183daa23d1" exitCode=0 Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.700929 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerDied","Data":"ff28f8e07e09ab9cd1123e1820ca7f55f0fdb1ecc1dd405fa046a7183daa23d1"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.701239 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerStarted","Data":"811985f126ca828f945fe10b5f3df79186434ed1336dd8b06f9a4a129b220a2b"} Dec 01 20:04:46 crc kubenswrapper[4960]: I1201 20:04:46.947165 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-tvqfm"] Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.713769 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tvqfm" event={"ID":"9445a1ff-2038-490e-9bf2-31d038d79ceb","Type":"ContainerStarted","Data":"65d842ea69cd7300ac4915f111b07672025c584563b99e88dafcac489c55f999"} Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.714200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tvqfm" event={"ID":"9445a1ff-2038-490e-9bf2-31d038d79ceb","Type":"ContainerStarted","Data":"eeaacce7503b2885fbc6b615b46deea726febffbd710dcf6603f006618be1ec6"} Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.716218 4960 generic.go:334] "Generic (PLEG): container finished" podID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerID="f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051" exitCode=0 Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.716276 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerDied","Data":"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051"} Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.720861 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerStarted","Data":"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446"} Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.742142 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-tvqfm" podStartSLOduration=2.742125609 podStartE2EDuration="2.742125609s" podCreationTimestamp="2025-12-01 20:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:47.739653659 +0000 UTC m=+1523.027145328" watchObservedRunningTime="2025-12-01 20:04:47.742125609 +0000 UTC m=+1523.029617288" Dec 01 20:04:47 crc kubenswrapper[4960]: I1201 20:04:47.770541 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.770510632 podStartE2EDuration="3.770510632s" podCreationTimestamp="2025-12-01 20:04:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:47.757426847 +0000 UTC m=+1523.044918516" watchObservedRunningTime="2025-12-01 20:04:47.770510632 +0000 UTC m=+1523.058002321" Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.017520 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.092292 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.092538 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="dnsmasq-dns" containerID="cri-o://1e4d4c2ddcc5c90dd903e202b5a38b04cefabdaab87df392c4190da1dd2b0ef7" gracePeriod=10 Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.736045 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerStarted","Data":"d04fed7483633c06890d66872983a3cf3122ed60a9748ac1de8cfcc692cccbbc"} Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.739159 4960 generic.go:334] "Generic (PLEG): container finished" podID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerID="1e4d4c2ddcc5c90dd903e202b5a38b04cefabdaab87df392c4190da1dd2b0ef7" exitCode=0 Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.739216 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" event={"ID":"04f34f8e-c023-476a-9a1e-f9c266fd1b86","Type":"ContainerDied","Data":"1e4d4c2ddcc5c90dd903e202b5a38b04cefabdaab87df392c4190da1dd2b0ef7"} Dec 01 20:04:48 crc kubenswrapper[4960]: I1201 20:04:48.746919 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerStarted","Data":"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c"} Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.314858 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.489445 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.489815 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.489939 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.490022 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.490059 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx4hf\" (UniqueName: \"kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.490291 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb\") pod \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\" (UID: \"04f34f8e-c023-476a-9a1e-f9c266fd1b86\") " Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.496754 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf" (OuterVolumeSpecName: "kube-api-access-xx4hf") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "kube-api-access-xx4hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.574077 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config" (OuterVolumeSpecName: "config") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.574792 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.579788 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.585685 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.587410 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "04f34f8e-c023-476a-9a1e-f9c266fd1b86" (UID: "04f34f8e-c023-476a-9a1e-f9c266fd1b86"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592554 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592574 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592583 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592595 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx4hf\" (UniqueName: \"kubernetes.io/projected/04f34f8e-c023-476a-9a1e-f9c266fd1b86-kube-api-access-xx4hf\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592603 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.592611 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/04f34f8e-c023-476a-9a1e-f9c266fd1b86-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.758177 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerStarted","Data":"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138"} Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.760483 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" event={"ID":"04f34f8e-c023-476a-9a1e-f9c266fd1b86","Type":"ContainerDied","Data":"fa39f48bcfcf4a97b398aeeb62b4fb76edb109ab27a6cae41441db66fd34174f"} Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.760509 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78cd565959-6tjw9" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.760541 4960 scope.go:117] "RemoveContainer" containerID="1e4d4c2ddcc5c90dd903e202b5a38b04cefabdaab87df392c4190da1dd2b0ef7" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.790358 4960 scope.go:117] "RemoveContainer" containerID="91f6896034eb54ed1ea0626a67f3fa5689f3c9ab41741b6e7db5788a56127b8c" Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.801989 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:04:49 crc kubenswrapper[4960]: I1201 20:04:49.810923 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78cd565959-6tjw9"] Dec 01 20:04:50 crc kubenswrapper[4960]: I1201 20:04:50.777390 4960 generic.go:334] "Generic (PLEG): container finished" podID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerID="d04fed7483633c06890d66872983a3cf3122ed60a9748ac1de8cfcc692cccbbc" exitCode=0 Dec 01 20:04:50 crc kubenswrapper[4960]: I1201 20:04:50.777487 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerDied","Data":"d04fed7483633c06890d66872983a3cf3122ed60a9748ac1de8cfcc692cccbbc"} Dec 01 20:04:50 crc kubenswrapper[4960]: I1201 20:04:50.786439 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerStarted","Data":"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e"} Dec 01 20:04:50 crc kubenswrapper[4960]: I1201 20:04:50.830988 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dq66q" podStartSLOduration=3.848448807 podStartE2EDuration="7.830960293s" podCreationTimestamp="2025-12-01 20:04:43 +0000 UTC" firstStartedPulling="2025-12-01 20:04:45.611105853 +0000 UTC m=+1520.898597522" lastFinishedPulling="2025-12-01 20:04:49.593617339 +0000 UTC m=+1524.881109008" observedRunningTime="2025-12-01 20:04:50.822752206 +0000 UTC m=+1526.110243875" watchObservedRunningTime="2025-12-01 20:04:50.830960293 +0000 UTC m=+1526.118451962" Dec 01 20:04:51 crc kubenswrapper[4960]: I1201 20:04:51.338364 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" path="/var/lib/kubelet/pods/04f34f8e-c023-476a-9a1e-f9c266fd1b86/volumes" Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.823356 4960 generic.go:334] "Generic (PLEG): container finished" podID="9445a1ff-2038-490e-9bf2-31d038d79ceb" containerID="65d842ea69cd7300ac4915f111b07672025c584563b99e88dafcac489c55f999" exitCode=0 Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.823469 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tvqfm" event={"ID":"9445a1ff-2038-490e-9bf2-31d038d79ceb","Type":"ContainerDied","Data":"65d842ea69cd7300ac4915f111b07672025c584563b99e88dafcac489c55f999"} Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.827043 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerStarted","Data":"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25"} Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.827419 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.830601 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerStarted","Data":"5fa51d98ab83a1a11058214ec066e19f95031fbc4be912242a13d584f741cfff"} Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.870876 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.08540224 podStartE2EDuration="9.870850976s" podCreationTimestamp="2025-12-01 20:04:43 +0000 UTC" firstStartedPulling="2025-12-01 20:04:45.538347507 +0000 UTC m=+1520.825839176" lastFinishedPulling="2025-12-01 20:04:52.323796243 +0000 UTC m=+1527.611287912" observedRunningTime="2025-12-01 20:04:52.864091727 +0000 UTC m=+1528.151583396" watchObservedRunningTime="2025-12-01 20:04:52.870850976 +0000 UTC m=+1528.158342645" Dec 01 20:04:52 crc kubenswrapper[4960]: I1201 20:04:52.891638 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdbr2" podStartSLOduration=3.269318369 podStartE2EDuration="8.891620522s" podCreationTimestamp="2025-12-01 20:04:44 +0000 UTC" firstStartedPulling="2025-12-01 20:04:46.702606085 +0000 UTC m=+1521.990097754" lastFinishedPulling="2025-12-01 20:04:52.324908238 +0000 UTC m=+1527.612399907" observedRunningTime="2025-12-01 20:04:52.884672835 +0000 UTC m=+1528.172164504" watchObservedRunningTime="2025-12-01 20:04:52.891620522 +0000 UTC m=+1528.179112191" Dec 01 20:04:53 crc kubenswrapper[4960]: I1201 20:04:53.784239 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:53 crc kubenswrapper[4960]: I1201 20:04:53.784558 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:53 crc kubenswrapper[4960]: I1201 20:04:53.872904 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.350901 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.499686 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts\") pod \"9445a1ff-2038-490e-9bf2-31d038d79ceb\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.499780 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle\") pod \"9445a1ff-2038-490e-9bf2-31d038d79ceb\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.499821 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9r6n\" (UniqueName: \"kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n\") pod \"9445a1ff-2038-490e-9bf2-31d038d79ceb\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.500015 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data\") pod \"9445a1ff-2038-490e-9bf2-31d038d79ceb\" (UID: \"9445a1ff-2038-490e-9bf2-31d038d79ceb\") " Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.506025 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts" (OuterVolumeSpecName: "scripts") pod "9445a1ff-2038-490e-9bf2-31d038d79ceb" (UID: "9445a1ff-2038-490e-9bf2-31d038d79ceb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.506333 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n" (OuterVolumeSpecName: "kube-api-access-h9r6n") pod "9445a1ff-2038-490e-9bf2-31d038d79ceb" (UID: "9445a1ff-2038-490e-9bf2-31d038d79ceb"). InnerVolumeSpecName "kube-api-access-h9r6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.542560 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9445a1ff-2038-490e-9bf2-31d038d79ceb" (UID: "9445a1ff-2038-490e-9bf2-31d038d79ceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.558695 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data" (OuterVolumeSpecName: "config-data") pod "9445a1ff-2038-490e-9bf2-31d038d79ceb" (UID: "9445a1ff-2038-490e-9bf2-31d038d79ceb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.602586 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.602626 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.602640 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9r6n\" (UniqueName: \"kubernetes.io/projected/9445a1ff-2038-490e-9bf2-31d038d79ceb-kube-api-access-h9r6n\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.602649 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9445a1ff-2038-490e-9bf2-31d038d79ceb-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.854298 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-tvqfm" event={"ID":"9445a1ff-2038-490e-9bf2-31d038d79ceb","Type":"ContainerDied","Data":"eeaacce7503b2885fbc6b615b46deea726febffbd710dcf6603f006618be1ec6"} Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.854341 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeaacce7503b2885fbc6b615b46deea726febffbd710dcf6603f006618be1ec6" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.854389 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-tvqfm" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.962503 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.966958 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 20:04:54 crc kubenswrapper[4960]: I1201 20:04:54.977328 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.044728 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.045004 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-log" containerID="cri-o://02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" gracePeriod=30 Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.045252 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-api" containerID="cri-o://1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" gracePeriod=30 Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.059311 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.059579 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" containerName="nova-scheduler-scheduler" containerID="cri-o://4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" gracePeriod=30 Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.097629 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.174216 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.174263 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.672246 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.681217 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.701620 4960 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.701700 4960 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" containerName="nova-scheduler-scheduler" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.869826 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.882542 4960 generic.go:334] "Generic (PLEG): container finished" podID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerID="1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" exitCode=0 Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.882803 4960 generic.go:334] "Generic (PLEG): container finished" podID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerID="02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" exitCode=143 Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.883217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerDied","Data":"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446"} Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.883275 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerDied","Data":"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2"} Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.883292 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d8f5a2af-5eb3-4209-b0b4-9c50afb13235","Type":"ContainerDied","Data":"14723fcd585a1fa936b8d9bd46aadaf9af9085601a65ff2f8e3847aa1c9c9c0c"} Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.883311 4960 scope.go:117] "RemoveContainer" containerID="1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.903631 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.934652 4960 scope.go:117] "RemoveContainer" containerID="02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.989149 4960 scope.go:117] "RemoveContainer" containerID="1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.991577 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446\": container with ID starting with 1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446 not found: ID does not exist" containerID="1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.991630 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446"} err="failed to get container status \"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446\": rpc error: code = NotFound desc = could not find container \"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446\": container with ID starting with 1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446 not found: ID does not exist" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.991659 4960 scope.go:117] "RemoveContainer" containerID="02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" Dec 01 20:04:55 crc kubenswrapper[4960]: E1201 20:04:55.991908 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2\": container with ID starting with 02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2 not found: ID does not exist" containerID="02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.991935 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2"} err="failed to get container status \"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2\": rpc error: code = NotFound desc = could not find container \"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2\": container with ID starting with 02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2 not found: ID does not exist" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.991952 4960 scope.go:117] "RemoveContainer" containerID="1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.992183 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446"} err="failed to get container status \"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446\": rpc error: code = NotFound desc = could not find container \"1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446\": container with ID starting with 1ed247ac692d5bb2002dcb353772e89ccc9cfb45d51b3ab306c692b36e9d5446 not found: ID does not exist" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.992209 4960 scope.go:117] "RemoveContainer" containerID="02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2" Dec 01 20:04:55 crc kubenswrapper[4960]: I1201 20:04:55.992401 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2"} err="failed to get container status \"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2\": rpc error: code = NotFound desc = could not find container \"02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2\": container with ID starting with 02874bf19043d724ecbea1c6ed5e102df7d06e772a30843672f80ee9d03ee3b2 not found: ID does not exist" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.032929 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsj9c\" (UniqueName: \"kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.032996 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.033063 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.033082 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.033160 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.033266 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs\") pod \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\" (UID: \"d8f5a2af-5eb3-4209-b0b4-9c50afb13235\") " Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.034572 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs" (OuterVolumeSpecName: "logs") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.043045 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c" (OuterVolumeSpecName: "kube-api-access-tsj9c") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "kube-api-access-tsj9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.080481 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data" (OuterVolumeSpecName: "config-data") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.097800 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.110989 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.111875 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d8f5a2af-5eb3-4209-b0b4-9c50afb13235" (UID: "d8f5a2af-5eb3-4209-b0b4-9c50afb13235"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135485 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135519 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135528 4960 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135539 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135549 4960 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.135559 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsj9c\" (UniqueName: \"kubernetes.io/projected/d8f5a2af-5eb3-4209-b0b4-9c50afb13235-kube-api-access-tsj9c\") on node \"crc\" DevicePath \"\"" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.223984 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bdbr2" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="registry-server" probeResult="failure" output=< Dec 01 20:04:56 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:04:56 crc kubenswrapper[4960]: > Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.896803 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.896859 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" containerID="cri-o://a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c" gracePeriod=30 Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.897927 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" containerID="cri-o://7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e" gracePeriod=30 Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.954596 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.967305 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.982445 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:56 crc kubenswrapper[4960]: E1201 20:04:56.984041 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="dnsmasq-dns" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984075 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="dnsmasq-dns" Dec 01 20:04:56 crc kubenswrapper[4960]: E1201 20:04:56.984131 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="init" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984148 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="init" Dec 01 20:04:56 crc kubenswrapper[4960]: E1201 20:04:56.984182 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9445a1ff-2038-490e-9bf2-31d038d79ceb" containerName="nova-manage" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984193 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9445a1ff-2038-490e-9bf2-31d038d79ceb" containerName="nova-manage" Dec 01 20:04:56 crc kubenswrapper[4960]: E1201 20:04:56.984229 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-api" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984241 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-api" Dec 01 20:04:56 crc kubenswrapper[4960]: E1201 20:04:56.984291 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-log" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984305 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-log" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.984976 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f34f8e-c023-476a-9a1e-f9c266fd1b86" containerName="dnsmasq-dns" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.985019 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-api" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.985036 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" containerName="nova-api-log" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.985043 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9445a1ff-2038-490e-9bf2-31d038d79ceb" containerName="nova-manage" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.987417 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.993634 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.994291 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 01 20:04:56 crc kubenswrapper[4960]: I1201 20:04:56.995168 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.025690 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057428 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-config-data\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057501 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb295cb-a487-406c-8f1d-517de7c1d245-logs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057525 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-public-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057580 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057634 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pltvx\" (UniqueName: \"kubernetes.io/projected/3cb295cb-a487-406c-8f1d-517de7c1d245-kube-api-access-pltvx\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.057667 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.159672 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-config-data\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.159785 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb295cb-a487-406c-8f1d-517de7c1d245-logs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.159809 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-public-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.160275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cb295cb-a487-406c-8f1d-517de7c1d245-logs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.160351 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.160679 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pltvx\" (UniqueName: \"kubernetes.io/projected/3cb295cb-a487-406c-8f1d-517de7c1d245-kube-api-access-pltvx\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.160719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.164889 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.165216 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.166448 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-config-data\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.169011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cb295cb-a487-406c-8f1d-517de7c1d245-public-tls-certs\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.176317 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pltvx\" (UniqueName: \"kubernetes.io/projected/3cb295cb-a487-406c-8f1d-517de7c1d245-kube-api-access-pltvx\") pod \"nova-api-0\" (UID: \"3cb295cb-a487-406c-8f1d-517de7c1d245\") " pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.335542 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f5a2af-5eb3-4209-b0b4-9c50afb13235" path="/var/lib/kubelet/pods/d8f5a2af-5eb3-4209-b0b4-9c50afb13235/volumes" Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.367505 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 01 20:04:57 crc kubenswrapper[4960]: W1201 20:04:57.862071 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cb295cb_a487_406c_8f1d_517de7c1d245.slice/crio-1dc48bfb3ec392fc79929913b256e29edd1a4ef63040c374b10ed795eed1865e WatchSource:0}: Error finding container 1dc48bfb3ec392fc79929913b256e29edd1a4ef63040c374b10ed795eed1865e: Status 404 returned error can't find the container with id 1dc48bfb3ec392fc79929913b256e29edd1a4ef63040c374b10ed795eed1865e Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.862928 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.916037 4960 generic.go:334] "Generic (PLEG): container finished" podID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerID="a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c" exitCode=143 Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.916203 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerDied","Data":"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c"} Dec 01 20:04:57 crc kubenswrapper[4960]: I1201 20:04:57.918057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cb295cb-a487-406c-8f1d-517de7c1d245","Type":"ContainerStarted","Data":"1dc48bfb3ec392fc79929913b256e29edd1a4ef63040c374b10ed795eed1865e"} Dec 01 20:04:58 crc kubenswrapper[4960]: I1201 20:04:58.931229 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cb295cb-a487-406c-8f1d-517de7c1d245","Type":"ContainerStarted","Data":"1deb5dd32ccf60cdc3af941e811145afe482f5ba828ba3c9e207b35b30954005"} Dec 01 20:04:58 crc kubenswrapper[4960]: I1201 20:04:58.931588 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3cb295cb-a487-406c-8f1d-517de7c1d245","Type":"ContainerStarted","Data":"ea9ea1eb89e5fa9dff606b298f5e5073d11a716f605f99d19b8f3bc4a012cdcb"} Dec 01 20:04:58 crc kubenswrapper[4960]: I1201 20:04:58.961436 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9614174220000002 podStartE2EDuration="2.961417422s" podCreationTimestamp="2025-12-01 20:04:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:04:58.952058227 +0000 UTC m=+1534.239549896" watchObservedRunningTime="2025-12-01 20:04:58.961417422 +0000 UTC m=+1534.248909091" Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.878148 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.943678 4960 generic.go:334] "Generic (PLEG): container finished" podID="20d77553-9219-4b2c-b0fc-5c642600a604" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" exitCode=0 Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.943764 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.943763 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"20d77553-9219-4b2c-b0fc-5c642600a604","Type":"ContainerDied","Data":"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599"} Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.943822 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"20d77553-9219-4b2c-b0fc-5c642600a604","Type":"ContainerDied","Data":"b82bd8250716f434241d163a5a316707a728867ebc25f070e6558ae64d097fee"} Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.943847 4960 scope.go:117] "RemoveContainer" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.971283 4960 scope.go:117] "RemoveContainer" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" Dec 01 20:04:59 crc kubenswrapper[4960]: E1201 20:04:59.971698 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599\": container with ID starting with 4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599 not found: ID does not exist" containerID="4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599" Dec 01 20:04:59 crc kubenswrapper[4960]: I1201 20:04:59.971744 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599"} err="failed to get container status \"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599\": rpc error: code = NotFound desc = could not find container \"4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599\": container with ID starting with 4fcb7ba6bf986381ba5c07c745d5879b48fd0202683327d6cc6c0f75c9fc3599 not found: ID does not exist" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.024131 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg6ls\" (UniqueName: \"kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls\") pod \"20d77553-9219-4b2c-b0fc-5c642600a604\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.024334 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle\") pod \"20d77553-9219-4b2c-b0fc-5c642600a604\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.024545 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data\") pod \"20d77553-9219-4b2c-b0fc-5c642600a604\" (UID: \"20d77553-9219-4b2c-b0fc-5c642600a604\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.042193 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": read tcp 10.217.0.2:56698->10.217.0.221:8775: read: connection reset by peer" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.042307 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.221:8775/\": read tcp 10.217.0.2:56702->10.217.0.221:8775: read: connection reset by peer" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.046692 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls" (OuterVolumeSpecName: "kube-api-access-mg6ls") pod "20d77553-9219-4b2c-b0fc-5c642600a604" (UID: "20d77553-9219-4b2c-b0fc-5c642600a604"). InnerVolumeSpecName "kube-api-access-mg6ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.059775 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20d77553-9219-4b2c-b0fc-5c642600a604" (UID: "20d77553-9219-4b2c-b0fc-5c642600a604"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.060272 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data" (OuterVolumeSpecName: "config-data") pod "20d77553-9219-4b2c-b0fc-5c642600a604" (UID: "20d77553-9219-4b2c-b0fc-5c642600a604"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.127084 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.127135 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg6ls\" (UniqueName: \"kubernetes.io/projected/20d77553-9219-4b2c-b0fc-5c642600a604-kube-api-access-mg6ls\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.127150 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d77553-9219-4b2c-b0fc-5c642600a604-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.283399 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.304472 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.318374 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:05:00 crc kubenswrapper[4960]: E1201 20:05:00.318937 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" containerName="nova-scheduler-scheduler" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.318969 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" containerName="nova-scheduler-scheduler" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.319334 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" containerName="nova-scheduler-scheduler" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.320360 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.324382 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.334764 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.431880 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.432167 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzfr4\" (UniqueName: \"kubernetes.io/projected/08cc519f-2c0d-444a-845a-9f403c9101be-kube-api-access-rzfr4\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.432227 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-config-data\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.533893 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzfr4\" (UniqueName: \"kubernetes.io/projected/08cc519f-2c0d-444a-845a-9f403c9101be-kube-api-access-rzfr4\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.533968 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-config-data\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.534091 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.539527 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-config-data\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.539822 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cc519f-2c0d-444a-845a-9f403c9101be-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.551681 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzfr4\" (UniqueName: \"kubernetes.io/projected/08cc519f-2c0d-444a-845a-9f403c9101be-kube-api-access-rzfr4\") pod \"nova-scheduler-0\" (UID: \"08cc519f-2c0d-444a-845a-9f403c9101be\") " pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.624168 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.645339 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.739907 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs\") pod \"319bb370-d760-40b2-89c2-e52d1ff260d5\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.739995 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data\") pod \"319bb370-d760-40b2-89c2-e52d1ff260d5\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.740185 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs\") pod \"319bb370-d760-40b2-89c2-e52d1ff260d5\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.740271 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z972m\" (UniqueName: \"kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m\") pod \"319bb370-d760-40b2-89c2-e52d1ff260d5\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.740300 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle\") pod \"319bb370-d760-40b2-89c2-e52d1ff260d5\" (UID: \"319bb370-d760-40b2-89c2-e52d1ff260d5\") " Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.741566 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs" (OuterVolumeSpecName: "logs") pod "319bb370-d760-40b2-89c2-e52d1ff260d5" (UID: "319bb370-d760-40b2-89c2-e52d1ff260d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.746846 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m" (OuterVolumeSpecName: "kube-api-access-z972m") pod "319bb370-d760-40b2-89c2-e52d1ff260d5" (UID: "319bb370-d760-40b2-89c2-e52d1ff260d5"). InnerVolumeSpecName "kube-api-access-z972m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.772652 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data" (OuterVolumeSpecName: "config-data") pod "319bb370-d760-40b2-89c2-e52d1ff260d5" (UID: "319bb370-d760-40b2-89c2-e52d1ff260d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.786839 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "319bb370-d760-40b2-89c2-e52d1ff260d5" (UID: "319bb370-d760-40b2-89c2-e52d1ff260d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.811911 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "319bb370-d760-40b2-89c2-e52d1ff260d5" (UID: "319bb370-d760-40b2-89c2-e52d1ff260d5"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.843465 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/319bb370-d760-40b2-89c2-e52d1ff260d5-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.843515 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.843549 4960 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.843565 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z972m\" (UniqueName: \"kubernetes.io/projected/319bb370-d760-40b2-89c2-e52d1ff260d5-kube-api-access-z972m\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.843578 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/319bb370-d760-40b2-89c2-e52d1ff260d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.957414 4960 generic.go:334] "Generic (PLEG): container finished" podID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerID="7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e" exitCode=0 Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.957457 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerDied","Data":"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e"} Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.957529 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"319bb370-d760-40b2-89c2-e52d1ff260d5","Type":"ContainerDied","Data":"d097f90f454e99aa850ed2e7fc7b03f551d8bfb35673eeb01a0f1317dd08c240"} Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.957550 4960 scope.go:117] "RemoveContainer" containerID="7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.957549 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:05:00 crc kubenswrapper[4960]: I1201 20:05:00.983653 4960 scope.go:117] "RemoveContainer" containerID="a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.025066 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.029668 4960 scope.go:117] "RemoveContainer" containerID="7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e" Dec 01 20:05:01 crc kubenswrapper[4960]: E1201 20:05:01.034747 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e\": container with ID starting with 7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e not found: ID does not exist" containerID="7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.034994 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e"} err="failed to get container status \"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e\": rpc error: code = NotFound desc = could not find container \"7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e\": container with ID starting with 7ff40c1a90d8c99660b6cb0a8b5fa59a63e1c9176456a4296514554b2bd8cd1e not found: ID does not exist" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.035091 4960 scope.go:117] "RemoveContainer" containerID="a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.037722 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: E1201 20:05:01.038985 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c\": container with ID starting with a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c not found: ID does not exist" containerID="a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.039034 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c"} err="failed to get container status \"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c\": rpc error: code = NotFound desc = could not find container \"a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c\": container with ID starting with a20989c18ab5c01a39c33bc7b0b462c11829f00350a0a7813cdf224b10b1dc4c not found: ID does not exist" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.053207 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: E1201 20:05:01.054016 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.054035 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" Dec 01 20:05:01 crc kubenswrapper[4960]: E1201 20:05:01.054054 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.054061 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.054254 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-log" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.054279 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" containerName="nova-metadata-metadata" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.055651 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.059052 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.059079 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.065589 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.150498 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.150586 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8tc\" (UniqueName: \"kubernetes.io/projected/feac0db7-ea9a-4add-9c01-f942bc15b1bf-kube-api-access-fv8tc\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.150630 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feac0db7-ea9a-4add-9c01-f942bc15b1bf-logs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.150674 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.150739 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-config-data\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.183989 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: W1201 20:05:01.184663 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08cc519f_2c0d_444a_845a_9f403c9101be.slice/crio-644c2b29c0d82e38e689f03d9f569340d1d7109ffca738959fe37a50e5361e1a WatchSource:0}: Error finding container 644c2b29c0d82e38e689f03d9f569340d1d7109ffca738959fe37a50e5361e1a: Status 404 returned error can't find the container with id 644c2b29c0d82e38e689f03d9f569340d1d7109ffca738959fe37a50e5361e1a Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.252580 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-config-data\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.252683 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.252739 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8tc\" (UniqueName: \"kubernetes.io/projected/feac0db7-ea9a-4add-9c01-f942bc15b1bf-kube-api-access-fv8tc\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.252775 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feac0db7-ea9a-4add-9c01-f942bc15b1bf-logs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.252821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.253539 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/feac0db7-ea9a-4add-9c01-f942bc15b1bf-logs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.256557 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.259090 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-config-data\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.259922 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/feac0db7-ea9a-4add-9c01-f942bc15b1bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.267490 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8tc\" (UniqueName: \"kubernetes.io/projected/feac0db7-ea9a-4add-9c01-f942bc15b1bf-kube-api-access-fv8tc\") pod \"nova-metadata-0\" (UID: \"feac0db7-ea9a-4add-9c01-f942bc15b1bf\") " pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.342592 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d77553-9219-4b2c-b0fc-5c642600a604" path="/var/lib/kubelet/pods/20d77553-9219-4b2c-b0fc-5c642600a604/volumes" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.343439 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="319bb370-d760-40b2-89c2-e52d1ff260d5" path="/var/lib/kubelet/pods/319bb370-d760-40b2-89c2-e52d1ff260d5/volumes" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.382635 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.832419 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.976630 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feac0db7-ea9a-4add-9c01-f942bc15b1bf","Type":"ContainerStarted","Data":"a86eaa3a972e112fcbbc71642c7aafbc9c0c920b0cb8c7b27cb4bf89d8ced45d"} Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.979627 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"08cc519f-2c0d-444a-845a-9f403c9101be","Type":"ContainerStarted","Data":"4a9c7eb0da25e1950ee9c34b69c7e075fc3d9994b8ce1541577e9ba3091ac59e"} Dec 01 20:05:01 crc kubenswrapper[4960]: I1201 20:05:01.979670 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"08cc519f-2c0d-444a-845a-9f403c9101be","Type":"ContainerStarted","Data":"644c2b29c0d82e38e689f03d9f569340d1d7109ffca738959fe37a50e5361e1a"} Dec 01 20:05:02 crc kubenswrapper[4960]: I1201 20:05:02.007463 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.007443583 podStartE2EDuration="2.007443583s" podCreationTimestamp="2025-12-01 20:05:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:05:02.000838458 +0000 UTC m=+1537.288330137" watchObservedRunningTime="2025-12-01 20:05:02.007443583 +0000 UTC m=+1537.294935242" Dec 01 20:05:02 crc kubenswrapper[4960]: I1201 20:05:02.997088 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feac0db7-ea9a-4add-9c01-f942bc15b1bf","Type":"ContainerStarted","Data":"353902d5c068692547be8f1a869a4bda2ef1795a587ec27d3f9bfdb633691488"} Dec 01 20:05:02 crc kubenswrapper[4960]: I1201 20:05:02.997418 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"feac0db7-ea9a-4add-9c01-f942bc15b1bf","Type":"ContainerStarted","Data":"8b63b4c1eadb3f8bc7266f86726b7227e451f52e068df2ce20786a3f7a3f62df"} Dec 01 20:05:03 crc kubenswrapper[4960]: I1201 20:05:03.019962 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.019942258 podStartE2EDuration="2.019942258s" podCreationTimestamp="2025-12-01 20:05:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:05:03.017945083 +0000 UTC m=+1538.305436772" watchObservedRunningTime="2025-12-01 20:05:03.019942258 +0000 UTC m=+1538.307433927" Dec 01 20:05:03 crc kubenswrapper[4960]: I1201 20:05:03.872035 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:05:03 crc kubenswrapper[4960]: I1201 20:05:03.934822 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.007622 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dq66q" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="registry-server" containerID="cri-o://8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e" gracePeriod=2 Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.605605 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.729505 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content\") pod \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.729656 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szh7x\" (UniqueName: \"kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x\") pod \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.729745 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities\") pod \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\" (UID: \"2c158cc1-23fd-4291-b3d1-ba55047c3f97\") " Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.730397 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities" (OuterVolumeSpecName: "utilities") pod "2c158cc1-23fd-4291-b3d1-ba55047c3f97" (UID: "2c158cc1-23fd-4291-b3d1-ba55047c3f97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.737433 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x" (OuterVolumeSpecName: "kube-api-access-szh7x") pod "2c158cc1-23fd-4291-b3d1-ba55047c3f97" (UID: "2c158cc1-23fd-4291-b3d1-ba55047c3f97"). InnerVolumeSpecName "kube-api-access-szh7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.784829 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c158cc1-23fd-4291-b3d1-ba55047c3f97" (UID: "2c158cc1-23fd-4291-b3d1-ba55047c3f97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.832941 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.832989 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szh7x\" (UniqueName: \"kubernetes.io/projected/2c158cc1-23fd-4291-b3d1-ba55047c3f97-kube-api-access-szh7x\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:04 crc kubenswrapper[4960]: I1201 20:05:04.833006 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c158cc1-23fd-4291-b3d1-ba55047c3f97-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.025735 4960 generic.go:334] "Generic (PLEG): container finished" podID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerID="8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e" exitCode=0 Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.025802 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerDied","Data":"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e"} Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.025867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dq66q" event={"ID":"2c158cc1-23fd-4291-b3d1-ba55047c3f97","Type":"ContainerDied","Data":"922e273e621f88ee3df9b7acb7f5ba1ae369595d1a36250c380b4d20684f9e57"} Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.025898 4960 scope.go:117] "RemoveContainer" containerID="8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.025804 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dq66q" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.058145 4960 scope.go:117] "RemoveContainer" containerID="f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.092325 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.105799 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dq66q"] Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.112288 4960 scope.go:117] "RemoveContainer" containerID="4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.168646 4960 scope.go:117] "RemoveContainer" containerID="8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e" Dec 01 20:05:05 crc kubenswrapper[4960]: E1201 20:05:05.169695 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e\": container with ID starting with 8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e not found: ID does not exist" containerID="8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.169740 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e"} err="failed to get container status \"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e\": rpc error: code = NotFound desc = could not find container \"8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e\": container with ID starting with 8521c37ff47bcc3f0ddf19dde55a76520b38cfaf864c8aec20643bff0db9815e not found: ID does not exist" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.169770 4960 scope.go:117] "RemoveContainer" containerID="f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051" Dec 01 20:05:05 crc kubenswrapper[4960]: E1201 20:05:05.170165 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051\": container with ID starting with f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051 not found: ID does not exist" containerID="f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.170223 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051"} err="failed to get container status \"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051\": rpc error: code = NotFound desc = could not find container \"f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051\": container with ID starting with f3f70b3d90d4919cc8e515b2633d0a359555f838dd52a7159cc037d18583c051 not found: ID does not exist" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.170261 4960 scope.go:117] "RemoveContainer" containerID="4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7" Dec 01 20:05:05 crc kubenswrapper[4960]: E1201 20:05:05.170779 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7\": container with ID starting with 4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7 not found: ID does not exist" containerID="4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.170813 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7"} err="failed to get container status \"4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7\": rpc error: code = NotFound desc = could not find container \"4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7\": container with ID starting with 4be3153222b4f5272527c26211bba384cf308d237758ca90373f8e2e99496ba7 not found: ID does not exist" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.233688 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.315987 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.343562 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" path="/var/lib/kubelet/pods/2c158cc1-23fd-4291-b3d1-ba55047c3f97/volumes" Dec 01 20:05:05 crc kubenswrapper[4960]: I1201 20:05:05.646032 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 01 20:05:06 crc kubenswrapper[4960]: I1201 20:05:06.383308 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:05:06 crc kubenswrapper[4960]: I1201 20:05:06.383593 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 01 20:05:07 crc kubenswrapper[4960]: I1201 20:05:07.368375 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:05:07 crc kubenswrapper[4960]: I1201 20:05:07.368444 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 01 20:05:07 crc kubenswrapper[4960]: I1201 20:05:07.531979 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:05:07 crc kubenswrapper[4960]: I1201 20:05:07.532337 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bdbr2" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="registry-server" containerID="cri-o://5fa51d98ab83a1a11058214ec066e19f95031fbc4be912242a13d584f741cfff" gracePeriod=2 Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.062917 4960 generic.go:334] "Generic (PLEG): container finished" podID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerID="5fa51d98ab83a1a11058214ec066e19f95031fbc4be912242a13d584f741cfff" exitCode=0 Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.063106 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerDied","Data":"5fa51d98ab83a1a11058214ec066e19f95031fbc4be912242a13d584f741cfff"} Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.063258 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdbr2" event={"ID":"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed","Type":"ContainerDied","Data":"811985f126ca828f945fe10b5f3df79186434ed1336dd8b06f9a4a129b220a2b"} Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.063280 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="811985f126ca828f945fe10b5f3df79186434ed1336dd8b06f9a4a129b220a2b" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.111908 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.211589 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities\") pod \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.211979 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4jdb\" (UniqueName: \"kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb\") pod \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.212172 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content\") pod \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\" (UID: \"c6fa8951-b9f2-4e80-90d0-fd5966fe42ed\") " Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.213881 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities" (OuterVolumeSpecName: "utilities") pod "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" (UID: "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.226708 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb" (OuterVolumeSpecName: "kube-api-access-x4jdb") pod "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" (UID: "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed"). InnerVolumeSpecName "kube-api-access-x4jdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.315150 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.315188 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4jdb\" (UniqueName: \"kubernetes.io/projected/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-kube-api-access-x4jdb\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.337512 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" (UID: "c6fa8951-b9f2-4e80-90d0-fd5966fe42ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.382358 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3cb295cb-a487-406c-8f1d-517de7c1d245" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.228:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.382669 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3cb295cb-a487-406c-8f1d-517de7c1d245" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.228:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:05:08 crc kubenswrapper[4960]: I1201 20:05:08.417819 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:09 crc kubenswrapper[4960]: I1201 20:05:09.076233 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdbr2" Dec 01 20:05:09 crc kubenswrapper[4960]: I1201 20:05:09.121723 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:05:09 crc kubenswrapper[4960]: I1201 20:05:09.134370 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bdbr2"] Dec 01 20:05:09 crc kubenswrapper[4960]: I1201 20:05:09.346738 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" path="/var/lib/kubelet/pods/c6fa8951-b9f2-4e80-90d0-fd5966fe42ed/volumes" Dec 01 20:05:10 crc kubenswrapper[4960]: I1201 20:05:10.646128 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 01 20:05:10 crc kubenswrapper[4960]: I1201 20:05:10.681383 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 01 20:05:11 crc kubenswrapper[4960]: I1201 20:05:11.160093 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 01 20:05:11 crc kubenswrapper[4960]: I1201 20:05:11.383406 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 20:05:11 crc kubenswrapper[4960]: I1201 20:05:11.383862 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 01 20:05:12 crc kubenswrapper[4960]: I1201 20:05:12.392902 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="feac0db7-ea9a-4add-9c01-f942bc15b1bf" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:05:12 crc kubenswrapper[4960]: I1201 20:05:12.392932 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="feac0db7-ea9a-4add-9c01-f942bc15b1bf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.000544 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.001826 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.001849 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.001876 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="extract-utilities" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.001887 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="extract-utilities" Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.001916 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="extract-content" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.001928 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="extract-content" Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.001961 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.001972 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.001998 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="extract-content" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.002008 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="extract-content" Dec 01 20:05:14 crc kubenswrapper[4960]: E1201 20:05:14.002037 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="extract-utilities" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.002047 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="extract-utilities" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.002410 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c158cc1-23fd-4291-b3d1-ba55047c3f97" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.002451 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6fa8951-b9f2-4e80-90d0-fd5966fe42ed" containerName="registry-server" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.004972 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.015533 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.043177 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.043225 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hntmp\" (UniqueName: \"kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.043244 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.145000 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.145069 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hntmp\" (UniqueName: \"kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.145092 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.145669 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.145948 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.174886 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hntmp\" (UniqueName: \"kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp\") pod \"community-operators-l2mg8\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.339644 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.378064 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 20:05:14 crc kubenswrapper[4960]: W1201 20:05:14.879402 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod358e5315_297e_4760_85ec_421b8114d10f.slice/crio-4c1699dca4524f1759aad49520fcc9712391e22f3c3955297501245eaf74349d WatchSource:0}: Error finding container 4c1699dca4524f1759aad49520fcc9712391e22f3c3955297501245eaf74349d: Status 404 returned error can't find the container with id 4c1699dca4524f1759aad49520fcc9712391e22f3c3955297501245eaf74349d Dec 01 20:05:14 crc kubenswrapper[4960]: I1201 20:05:14.883490 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:15 crc kubenswrapper[4960]: I1201 20:05:15.158853 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerStarted","Data":"4c1699dca4524f1759aad49520fcc9712391e22f3c3955297501245eaf74349d"} Dec 01 20:05:16 crc kubenswrapper[4960]: I1201 20:05:16.172288 4960 generic.go:334] "Generic (PLEG): container finished" podID="358e5315-297e-4760-85ec-421b8114d10f" containerID="ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0" exitCode=0 Dec 01 20:05:16 crc kubenswrapper[4960]: I1201 20:05:16.172350 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerDied","Data":"ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0"} Dec 01 20:05:17 crc kubenswrapper[4960]: I1201 20:05:17.375746 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 20:05:17 crc kubenswrapper[4960]: I1201 20:05:17.376452 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 20:05:17 crc kubenswrapper[4960]: I1201 20:05:17.378958 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 01 20:05:17 crc kubenswrapper[4960]: I1201 20:05:17.387697 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 20:05:18 crc kubenswrapper[4960]: I1201 20:05:18.199575 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerStarted","Data":"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a"} Dec 01 20:05:18 crc kubenswrapper[4960]: I1201 20:05:18.200075 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 01 20:05:18 crc kubenswrapper[4960]: I1201 20:05:18.211140 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 01 20:05:19 crc kubenswrapper[4960]: I1201 20:05:19.215907 4960 generic.go:334] "Generic (PLEG): container finished" podID="358e5315-297e-4760-85ec-421b8114d10f" containerID="ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a" exitCode=0 Dec 01 20:05:19 crc kubenswrapper[4960]: I1201 20:05:19.216001 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerDied","Data":"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a"} Dec 01 20:05:20 crc kubenswrapper[4960]: I1201 20:05:20.231682 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerStarted","Data":"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765"} Dec 01 20:05:21 crc kubenswrapper[4960]: I1201 20:05:21.391191 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 20:05:21 crc kubenswrapper[4960]: I1201 20:05:21.393717 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 01 20:05:21 crc kubenswrapper[4960]: I1201 20:05:21.400468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 20:05:21 crc kubenswrapper[4960]: I1201 20:05:21.402422 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 01 20:05:21 crc kubenswrapper[4960]: I1201 20:05:21.417012 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l2mg8" podStartSLOduration=4.865305745 podStartE2EDuration="8.416991635s" podCreationTimestamp="2025-12-01 20:05:13 +0000 UTC" firstStartedPulling="2025-12-01 20:05:16.174388729 +0000 UTC m=+1551.461880438" lastFinishedPulling="2025-12-01 20:05:19.726074649 +0000 UTC m=+1555.013566328" observedRunningTime="2025-12-01 20:05:20.257533644 +0000 UTC m=+1555.545025383" watchObservedRunningTime="2025-12-01 20:05:21.416991635 +0000 UTC m=+1556.704483304" Dec 01 20:05:24 crc kubenswrapper[4960]: I1201 20:05:24.340516 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:24 crc kubenswrapper[4960]: I1201 20:05:24.341177 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:24 crc kubenswrapper[4960]: I1201 20:05:24.396705 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:25 crc kubenswrapper[4960]: I1201 20:05:25.351433 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:25 crc kubenswrapper[4960]: I1201 20:05:25.413347 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:27 crc kubenswrapper[4960]: I1201 20:05:27.318658 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l2mg8" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="registry-server" containerID="cri-o://157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765" gracePeriod=2 Dec 01 20:05:27 crc kubenswrapper[4960]: I1201 20:05:27.948052 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.052997 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content\") pod \"358e5315-297e-4760-85ec-421b8114d10f\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.053178 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities\") pod \"358e5315-297e-4760-85ec-421b8114d10f\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.053334 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hntmp\" (UniqueName: \"kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp\") pod \"358e5315-297e-4760-85ec-421b8114d10f\" (UID: \"358e5315-297e-4760-85ec-421b8114d10f\") " Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.054529 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities" (OuterVolumeSpecName: "utilities") pod "358e5315-297e-4760-85ec-421b8114d10f" (UID: "358e5315-297e-4760-85ec-421b8114d10f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.062311 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp" (OuterVolumeSpecName: "kube-api-access-hntmp") pod "358e5315-297e-4760-85ec-421b8114d10f" (UID: "358e5315-297e-4760-85ec-421b8114d10f"). InnerVolumeSpecName "kube-api-access-hntmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.114182 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "358e5315-297e-4760-85ec-421b8114d10f" (UID: "358e5315-297e-4760-85ec-421b8114d10f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.156172 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.156203 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hntmp\" (UniqueName: \"kubernetes.io/projected/358e5315-297e-4760-85ec-421b8114d10f-kube-api-access-hntmp\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.156213 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/358e5315-297e-4760-85ec-421b8114d10f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.336530 4960 generic.go:334] "Generic (PLEG): container finished" podID="358e5315-297e-4760-85ec-421b8114d10f" containerID="157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765" exitCode=0 Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.336576 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerDied","Data":"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765"} Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.336606 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2mg8" event={"ID":"358e5315-297e-4760-85ec-421b8114d10f","Type":"ContainerDied","Data":"4c1699dca4524f1759aad49520fcc9712391e22f3c3955297501245eaf74349d"} Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.336622 4960 scope.go:117] "RemoveContainer" containerID="157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.336683 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2mg8" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.365290 4960 scope.go:117] "RemoveContainer" containerID="ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.425345 4960 scope.go:117] "RemoveContainer" containerID="ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.426329 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.442097 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l2mg8"] Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.475102 4960 scope.go:117] "RemoveContainer" containerID="157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765" Dec 01 20:05:28 crc kubenswrapper[4960]: E1201 20:05:28.475573 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765\": container with ID starting with 157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765 not found: ID does not exist" containerID="157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.475861 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765"} err="failed to get container status \"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765\": rpc error: code = NotFound desc = could not find container \"157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765\": container with ID starting with 157165171b9f7c36a5ed6eaa904d2d4106586fed08a9be1db050bba461f01765 not found: ID does not exist" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.475976 4960 scope.go:117] "RemoveContainer" containerID="ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a" Dec 01 20:05:28 crc kubenswrapper[4960]: E1201 20:05:28.476604 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a\": container with ID starting with ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a not found: ID does not exist" containerID="ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.476650 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a"} err="failed to get container status \"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a\": rpc error: code = NotFound desc = could not find container \"ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a\": container with ID starting with ae54f13442b6eb9e89ae3f8a13030bdcffb2bd1976d50e8cba4ee7e95d8a6e8a not found: ID does not exist" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.476681 4960 scope.go:117] "RemoveContainer" containerID="ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0" Dec 01 20:05:28 crc kubenswrapper[4960]: E1201 20:05:28.476917 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0\": container with ID starting with ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0 not found: ID does not exist" containerID="ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0" Dec 01 20:05:28 crc kubenswrapper[4960]: I1201 20:05:28.476939 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0"} err="failed to get container status \"ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0\": rpc error: code = NotFound desc = could not find container \"ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0\": container with ID starting with ef664e8c34cd9c28c62eeea09b9401b3d20e6f7ded252879d112cb04f94461a0 not found: ID does not exist" Dec 01 20:05:29 crc kubenswrapper[4960]: I1201 20:05:29.342316 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="358e5315-297e-4760-85ec-421b8114d10f" path="/var/lib/kubelet/pods/358e5315-297e-4760-85ec-421b8114d10f/volumes" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.397552 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-zdrn9"] Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.408308 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-zdrn9"] Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.503782 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-db-sync-dl272"] Dec 01 20:05:32 crc kubenswrapper[4960]: E1201 20:05:32.504378 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="registry-server" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.504398 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="registry-server" Dec 01 20:05:32 crc kubenswrapper[4960]: E1201 20:05:32.504468 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="extract-content" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.504478 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="extract-content" Dec 01 20:05:32 crc kubenswrapper[4960]: E1201 20:05:32.504491 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="extract-utilities" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.504499 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="extract-utilities" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.504737 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="358e5315-297e-4760-85ec-421b8114d10f" containerName="registry-server" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.505719 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.507657 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.515681 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-dl272"] Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.651766 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.652488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.652534 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.652648 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.652763 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p4cm\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.755256 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p4cm\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.755391 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.755430 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.755459 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.755531 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.761293 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.762654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.763450 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.764841 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.773798 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p4cm\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm\") pod \"cloudkitty-db-sync-dl272\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:32 crc kubenswrapper[4960]: I1201 20:05:32.828023 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:33 crc kubenswrapper[4960]: I1201 20:05:33.319998 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-db-sync-dl272"] Dec 01 20:05:33 crc kubenswrapper[4960]: I1201 20:05:33.335789 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b90b04-9bf0-4bab-a581-4691ef53713e" path="/var/lib/kubelet/pods/80b90b04-9bf0-4bab-a581-4691ef53713e/volumes" Dec 01 20:05:33 crc kubenswrapper[4960]: I1201 20:05:33.395456 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-dl272" event={"ID":"f838587c-b056-4d32-bf13-e6a7a0c8da50","Type":"ContainerStarted","Data":"a4586bf3cb31f30c4ed6e9e72e763f8f314ad4f46206a8c98ba0539d25c7ec75"} Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.031327 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.032381 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-central-agent" containerID="cri-o://a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511" gracePeriod=30 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.032451 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="proxy-httpd" containerID="cri-o://363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25" gracePeriod=30 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.032525 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="sg-core" containerID="cri-o://a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138" gracePeriod=30 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.032586 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-notification-agent" containerID="cri-o://f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c" gracePeriod=30 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.407775 4960 generic.go:334] "Generic (PLEG): container finished" podID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerID="363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25" exitCode=0 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.408157 4960 generic.go:334] "Generic (PLEG): container finished" podID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerID="a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138" exitCode=2 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.408215 4960 generic.go:334] "Generic (PLEG): container finished" podID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerID="a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511" exitCode=0 Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.407867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerDied","Data":"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25"} Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.408386 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerDied","Data":"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138"} Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.408462 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerDied","Data":"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511"} Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.410321 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-dl272" event={"ID":"f838587c-b056-4d32-bf13-e6a7a0c8da50","Type":"ContainerStarted","Data":"da51dfbd41632d77feaf2e9ae1b58baa07d8556b44d96181be1f69390f4bc9c0"} Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.428571 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-db-sync-dl272" podStartSLOduration=2.241702654 podStartE2EDuration="2.428554325s" podCreationTimestamp="2025-12-01 20:05:32 +0000 UTC" firstStartedPulling="2025-12-01 20:05:33.319638779 +0000 UTC m=+1568.607130458" lastFinishedPulling="2025-12-01 20:05:33.50649046 +0000 UTC m=+1568.793982129" observedRunningTime="2025-12-01 20:05:34.426036187 +0000 UTC m=+1569.713527856" watchObservedRunningTime="2025-12-01 20:05:34.428554325 +0000 UTC m=+1569.716045994" Dec 01 20:05:34 crc kubenswrapper[4960]: I1201 20:05:34.821039 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:35 crc kubenswrapper[4960]: I1201 20:05:35.631818 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.114893 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.232608 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.232678 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.232770 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.232836 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.232959 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.233066 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmnjk\" (UniqueName: \"kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.233092 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.233168 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts\") pod \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\" (UID: \"76c6a0f6-dfa9-41eb-9909-d1a9417d5909\") " Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.235385 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.237724 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.264245 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts" (OuterVolumeSpecName: "scripts") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.264305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk" (OuterVolumeSpecName: "kube-api-access-xmnjk") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "kube-api-access-xmnjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.304944 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.309305 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337529 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337561 4960 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337571 4960 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337580 4960 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337588 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.337599 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmnjk\" (UniqueName: \"kubernetes.io/projected/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-kube-api-access-xmnjk\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.344810 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.372291 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data" (OuterVolumeSpecName: "config-data") pod "76c6a0f6-dfa9-41eb-9909-d1a9417d5909" (UID: "76c6a0f6-dfa9-41eb-9909-d1a9417d5909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.435899 4960 generic.go:334] "Generic (PLEG): container finished" podID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerID="f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c" exitCode=0 Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.436022 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerDied","Data":"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c"} Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.436080 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76c6a0f6-dfa9-41eb-9909-d1a9417d5909","Type":"ContainerDied","Data":"ea44aaee72053234872a11ce12e4e1605850ae5de7541d74ccb63e387e78dfb1"} Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.436098 4960 scope.go:117] "RemoveContainer" containerID="363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.436453 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.437815 4960 generic.go:334] "Generic (PLEG): container finished" podID="f838587c-b056-4d32-bf13-e6a7a0c8da50" containerID="da51dfbd41632d77feaf2e9ae1b58baa07d8556b44d96181be1f69390f4bc9c0" exitCode=0 Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.437850 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-dl272" event={"ID":"f838587c-b056-4d32-bf13-e6a7a0c8da50","Type":"ContainerDied","Data":"da51dfbd41632d77feaf2e9ae1b58baa07d8556b44d96181be1f69390f4bc9c0"} Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.439033 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.439053 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76c6a0f6-dfa9-41eb-9909-d1a9417d5909-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.471104 4960 scope.go:117] "RemoveContainer" containerID="a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.492638 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.496267 4960 scope.go:117] "RemoveContainer" containerID="f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.506236 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.515649 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.516100 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="sg-core" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.516128 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="sg-core" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.516144 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-central-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.516150 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-central-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.517164 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-notification-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517182 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-notification-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.517194 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="proxy-httpd" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517211 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="proxy-httpd" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517424 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="proxy-httpd" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517442 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="sg-core" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517463 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-central-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.517482 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" containerName="ceilometer-notification-agent" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.519364 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.523155 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.523341 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.523595 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.527377 4960 scope.go:117] "RemoveContainer" containerID="a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.533348 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.563742 4960 scope.go:117] "RemoveContainer" containerID="363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.564360 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25\": container with ID starting with 363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25 not found: ID does not exist" containerID="363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.564388 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25"} err="failed to get container status \"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25\": rpc error: code = NotFound desc = could not find container \"363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25\": container with ID starting with 363ca4b586b0d2486861eeee586242e0d9e06fc6b6166624c151b4555587cf25 not found: ID does not exist" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.564410 4960 scope.go:117] "RemoveContainer" containerID="a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.565206 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138\": container with ID starting with a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138 not found: ID does not exist" containerID="a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.565269 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138"} err="failed to get container status \"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138\": rpc error: code = NotFound desc = could not find container \"a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138\": container with ID starting with a9a7c98a1cda3188e4382f4fb25fea3f37077deacc638d88b682caa74a1a3138 not found: ID does not exist" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.565306 4960 scope.go:117] "RemoveContainer" containerID="f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.566649 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c\": container with ID starting with f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c not found: ID does not exist" containerID="f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.566697 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c"} err="failed to get container status \"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c\": rpc error: code = NotFound desc = could not find container \"f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c\": container with ID starting with f86d1021a2ddc8821b8e4d91bea3a5e3f5292d6e82e424be5782aa0edcfbc31c not found: ID does not exist" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.566729 4960 scope.go:117] "RemoveContainer" containerID="a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511" Dec 01 20:05:36 crc kubenswrapper[4960]: E1201 20:05:36.567202 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511\": container with ID starting with a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511 not found: ID does not exist" containerID="a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.567246 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511"} err="failed to get container status \"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511\": rpc error: code = NotFound desc = could not find container \"a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511\": container with ID starting with a3c5a083a5e2322cdf75a6780328851603ec239393ad2fbe507073370a90f511 not found: ID does not exist" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.642821 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.642873 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-log-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.642957 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-config-data\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.643044 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-scripts\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.643193 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.643296 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.643378 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vzlj\" (UniqueName: \"kubernetes.io/projected/62742f95-55dd-471c-964f-33b6eea63418-kube-api-access-7vzlj\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.643536 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-run-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-config-data\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745474 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-scripts\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745508 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745537 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745568 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vzlj\" (UniqueName: \"kubernetes.io/projected/62742f95-55dd-471c-964f-33b6eea63418-kube-api-access-7vzlj\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745612 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-run-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745648 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.745666 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-log-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.746147 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-log-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.746396 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62742f95-55dd-471c-964f-33b6eea63418-run-httpd\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.749324 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-scripts\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.749875 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.750041 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-config-data\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.750438 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.750625 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62742f95-55dd-471c-964f-33b6eea63418-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.764727 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vzlj\" (UniqueName: \"kubernetes.io/projected/62742f95-55dd-471c-964f-33b6eea63418-kube-api-access-7vzlj\") pod \"ceilometer-0\" (UID: \"62742f95-55dd-471c-964f-33b6eea63418\") " pod="openstack/ceilometer-0" Dec 01 20:05:36 crc kubenswrapper[4960]: I1201 20:05:36.848455 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.335992 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76c6a0f6-dfa9-41eb-9909-d1a9417d5909" path="/var/lib/kubelet/pods/76c6a0f6-dfa9-41eb-9909-d1a9417d5909/volumes" Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.369094 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.450805 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62742f95-55dd-471c-964f-33b6eea63418","Type":"ContainerStarted","Data":"77f175923dc4805bac9738bbc0bc1e474a62c9fac68ceb52e7e5ae85737774e0"} Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.851366 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.969618 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data\") pod \"f838587c-b056-4d32-bf13-e6a7a0c8da50\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.970031 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle\") pod \"f838587c-b056-4d32-bf13-e6a7a0c8da50\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.970072 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs\") pod \"f838587c-b056-4d32-bf13-e6a7a0c8da50\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.970155 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p4cm\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm\") pod \"f838587c-b056-4d32-bf13-e6a7a0c8da50\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.970363 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts\") pod \"f838587c-b056-4d32-bf13-e6a7a0c8da50\" (UID: \"f838587c-b056-4d32-bf13-e6a7a0c8da50\") " Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.977275 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts" (OuterVolumeSpecName: "scripts") pod "f838587c-b056-4d32-bf13-e6a7a0c8da50" (UID: "f838587c-b056-4d32-bf13-e6a7a0c8da50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.992366 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs" (OuterVolumeSpecName: "certs") pod "f838587c-b056-4d32-bf13-e6a7a0c8da50" (UID: "f838587c-b056-4d32-bf13-e6a7a0c8da50"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:37 crc kubenswrapper[4960]: I1201 20:05:37.994324 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm" (OuterVolumeSpecName: "kube-api-access-6p4cm") pod "f838587c-b056-4d32-bf13-e6a7a0c8da50" (UID: "f838587c-b056-4d32-bf13-e6a7a0c8da50"). InnerVolumeSpecName "kube-api-access-6p4cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.013577 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f838587c-b056-4d32-bf13-e6a7a0c8da50" (UID: "f838587c-b056-4d32-bf13-e6a7a0c8da50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.020258 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data" (OuterVolumeSpecName: "config-data") pod "f838587c-b056-4d32-bf13-e6a7a0c8da50" (UID: "f838587c-b056-4d32-bf13-e6a7a0c8da50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.074201 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.074228 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.074238 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.074248 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p4cm\" (UniqueName: \"kubernetes.io/projected/f838587c-b056-4d32-bf13-e6a7a0c8da50-kube-api-access-6p4cm\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.074257 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f838587c-b056-4d32-bf13-e6a7a0c8da50-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.463291 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-db-sync-dl272" event={"ID":"f838587c-b056-4d32-bf13-e6a7a0c8da50","Type":"ContainerDied","Data":"a4586bf3cb31f30c4ed6e9e72e763f8f314ad4f46206a8c98ba0539d25c7ec75"} Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.463328 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4586bf3cb31f30c4ed6e9e72e763f8f314ad4f46206a8c98ba0539d25c7ec75" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.463363 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-db-sync-dl272" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.537646 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-8vlfj"] Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.551996 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-8vlfj"] Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.654176 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-storageinit-bsnf5"] Dec 01 20:05:38 crc kubenswrapper[4960]: E1201 20:05:38.654588 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f838587c-b056-4d32-bf13-e6a7a0c8da50" containerName="cloudkitty-db-sync" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.654604 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f838587c-b056-4d32-bf13-e6a7a0c8da50" containerName="cloudkitty-db-sync" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.654815 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f838587c-b056-4d32-bf13-e6a7a0c8da50" containerName="cloudkitty-db-sync" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.655542 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.672368 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.686702 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-bsnf5"] Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.789839 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.789889 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.789917 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.790171 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cr9g\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.790293 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.892422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.892791 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.892874 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cr9g\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.892920 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.893028 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.908024 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.915119 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.915466 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.929975 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cr9g\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.931312 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data\") pod \"cloudkitty-storageinit-bsnf5\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:38 crc kubenswrapper[4960]: I1201 20:05:38.973664 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:39 crc kubenswrapper[4960]: I1201 20:05:39.340383 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62a5e00a-99c2-4caa-a683-243167f19130" path="/var/lib/kubelet/pods/62a5e00a-99c2-4caa-a683-243167f19130/volumes" Dec 01 20:05:39 crc kubenswrapper[4960]: I1201 20:05:39.561420 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-storageinit-bsnf5"] Dec 01 20:05:39 crc kubenswrapper[4960]: I1201 20:05:39.874785 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="rabbitmq" containerID="cri-o://3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4" gracePeriod=604795 Dec 01 20:05:40 crc kubenswrapper[4960]: I1201 20:05:40.211521 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="rabbitmq" containerID="cri-o://85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a" gracePeriod=604796 Dec 01 20:05:41 crc kubenswrapper[4960]: W1201 20:05:41.415917 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c330c49_1896_4917_a0cf_4d5d5f59f25c.slice/crio-3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b WatchSource:0}: Error finding container 3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b: Status 404 returned error can't find the container with id 3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b Dec 01 20:05:41 crc kubenswrapper[4960]: I1201 20:05:41.500867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-bsnf5" event={"ID":"4c330c49-1896-4917-a0cf-4d5d5f59f25c","Type":"ContainerStarted","Data":"3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b"} Dec 01 20:05:42 crc kubenswrapper[4960]: I1201 20:05:42.512842 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-bsnf5" event={"ID":"4c330c49-1896-4917-a0cf-4d5d5f59f25c","Type":"ContainerStarted","Data":"12ea03fb40e20df4f24a4d5c3267e8540db8e91cc9301298b637e189264ff046"} Dec 01 20:05:42 crc kubenswrapper[4960]: I1201 20:05:42.515179 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62742f95-55dd-471c-964f-33b6eea63418","Type":"ContainerStarted","Data":"ebb7d8e03e642f4b42ea09666c8bfbc4bb343fdc6e3889d09dfec19d5566eff6"} Dec 01 20:05:42 crc kubenswrapper[4960]: I1201 20:05:42.534723 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-storageinit-bsnf5" podStartSLOduration=4.534698507 podStartE2EDuration="4.534698507s" podCreationTimestamp="2025-12-01 20:05:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:05:42.527860334 +0000 UTC m=+1577.815352003" watchObservedRunningTime="2025-12-01 20:05:42.534698507 +0000 UTC m=+1577.822190176" Dec 01 20:05:43 crc kubenswrapper[4960]: I1201 20:05:43.527740 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c330c49-1896-4917-a0cf-4d5d5f59f25c" containerID="12ea03fb40e20df4f24a4d5c3267e8540db8e91cc9301298b637e189264ff046" exitCode=0 Dec 01 20:05:43 crc kubenswrapper[4960]: I1201 20:05:43.527980 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-bsnf5" event={"ID":"4c330c49-1896-4917-a0cf-4d5d5f59f25c","Type":"ContainerDied","Data":"12ea03fb40e20df4f24a4d5c3267e8540db8e91cc9301298b637e189264ff046"} Dec 01 20:05:43 crc kubenswrapper[4960]: I1201 20:05:43.530642 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62742f95-55dd-471c-964f-33b6eea63418","Type":"ContainerStarted","Data":"8e507ef95e56f6d8917173ac3b0ecf8e9fad3f360b05ef6b77a67100929cb155"} Dec 01 20:05:44 crc kubenswrapper[4960]: I1201 20:05:44.350357 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Dec 01 20:05:44 crc kubenswrapper[4960]: I1201 20:05:44.467243 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Dec 01 20:05:44 crc kubenswrapper[4960]: I1201 20:05:44.543916 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62742f95-55dd-471c-964f-33b6eea63418","Type":"ContainerStarted","Data":"d0c2d27e62fc2ae1e16b4eff8733398e4c4fcb009197ae4c0a3e208228dc330b"} Dec 01 20:05:44 crc kubenswrapper[4960]: I1201 20:05:44.971231 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.122021 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cr9g\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g\") pod \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.122207 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs\") pod \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.122240 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle\") pod \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.122271 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts\") pod \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.122389 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data\") pod \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\" (UID: \"4c330c49-1896-4917-a0cf-4d5d5f59f25c\") " Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.130018 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts" (OuterVolumeSpecName: "scripts") pod "4c330c49-1896-4917-a0cf-4d5d5f59f25c" (UID: "4c330c49-1896-4917-a0cf-4d5d5f59f25c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.130832 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g" (OuterVolumeSpecName: "kube-api-access-6cr9g") pod "4c330c49-1896-4917-a0cf-4d5d5f59f25c" (UID: "4c330c49-1896-4917-a0cf-4d5d5f59f25c"). InnerVolumeSpecName "kube-api-access-6cr9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.133166 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs" (OuterVolumeSpecName: "certs") pod "4c330c49-1896-4917-a0cf-4d5d5f59f25c" (UID: "4c330c49-1896-4917-a0cf-4d5d5f59f25c"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.163786 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c330c49-1896-4917-a0cf-4d5d5f59f25c" (UID: "4c330c49-1896-4917-a0cf-4d5d5f59f25c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.166972 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data" (OuterVolumeSpecName: "config-data") pod "4c330c49-1896-4917-a0cf-4d5d5f59f25c" (UID: "4c330c49-1896-4917-a0cf-4d5d5f59f25c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.225671 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.225712 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.225728 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.225739 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c330c49-1896-4917-a0cf-4d5d5f59f25c-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.225752 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cr9g\" (UniqueName: \"kubernetes.io/projected/4c330c49-1896-4917-a0cf-4d5d5f59f25c-kube-api-access-6cr9g\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.559781 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-storageinit-bsnf5" event={"ID":"4c330c49-1896-4917-a0cf-4d5d5f59f25c","Type":"ContainerDied","Data":"3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b"} Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.560263 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f18fdda4ab0d488076346e4c26bbca07220c64ea9b8d2416d9af67d2702003b" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.559837 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-storageinit-bsnf5" Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.703916 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.704140 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-proc-0" podUID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" containerName="cloudkitty-proc" containerID="cri-o://7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3" gracePeriod=30 Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.718304 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.718586 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api-log" containerID="cri-o://29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a" gracePeriod=30 Dec 01 20:05:45 crc kubenswrapper[4960]: I1201 20:05:45.719133 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cloudkitty-api-0" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api" containerID="cri-o://6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac" gracePeriod=30 Dec 01 20:05:46 crc kubenswrapper[4960]: I1201 20:05:46.578380 4960 generic.go:334] "Generic (PLEG): container finished" podID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerID="29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a" exitCode=143 Dec 01 20:05:46 crc kubenswrapper[4960]: I1201 20:05:46.578439 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerDied","Data":"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a"} Dec 01 20:05:46 crc kubenswrapper[4960]: I1201 20:05:46.583912 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62742f95-55dd-471c-964f-33b6eea63418","Type":"ContainerStarted","Data":"63976896d03674e0c8dc5b1610357547de3d7081b9663e5b666a5fed509bc6c5"} Dec 01 20:05:46 crc kubenswrapper[4960]: I1201 20:05:46.584577 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 01 20:05:46 crc kubenswrapper[4960]: I1201 20:05:46.608778 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.320779851 podStartE2EDuration="10.608757916s" podCreationTimestamp="2025-12-01 20:05:36 +0000 UTC" firstStartedPulling="2025-12-01 20:05:37.377082432 +0000 UTC m=+1572.664574101" lastFinishedPulling="2025-12-01 20:05:45.665060507 +0000 UTC m=+1580.952552166" observedRunningTime="2025-12-01 20:05:46.604349949 +0000 UTC m=+1581.891841618" watchObservedRunningTime="2025-12-01 20:05:46.608757916 +0000 UTC m=+1581.896249585" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.232251 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379413 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379510 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379559 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379616 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379703 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zds2p\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.379738 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data\") pod \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\" (UID: \"21d1f29c-ee4b-4009-8017-ee502c3cfab9\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.385613 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts" (OuterVolumeSpecName: "scripts") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.385932 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p" (OuterVolumeSpecName: "kube-api-access-zds2p") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "kube-api-access-zds2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.389201 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.391399 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs" (OuterVolumeSpecName: "certs") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.424408 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.427852 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data" (OuterVolumeSpecName: "config-data") pod "21d1f29c-ee4b-4009-8017-ee502c3cfab9" (UID: "21d1f29c-ee4b-4009-8017-ee502c3cfab9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482593 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482632 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482642 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482650 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zds2p\" (UniqueName: \"kubernetes.io/projected/21d1f29c-ee4b-4009-8017-ee502c3cfab9-kube-api-access-zds2p\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482658 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.482666 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21d1f29c-ee4b-4009-8017-ee502c3cfab9-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.485679 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.492776 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.501005 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584281 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584333 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584372 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584420 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584447 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584470 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xssk8\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584507 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584537 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.584562 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585218 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585535 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585573 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585602 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585625 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585660 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585691 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585722 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrpz6\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585765 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585801 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585788 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.585885 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587023 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587046 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587166 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587205 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587303 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587323 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587366 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587390 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2shq5\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587408 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587436 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data\") pod \"c2b84514-def8-482d-a4a3-8b71808fcdcc\" (UID: \"c2b84514-def8-482d-a4a3-8b71808fcdcc\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587464 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls\") pod \"87ca6976-ad61-40c2-97a4-56ca46df7448\" (UID: \"87ca6976-ad61-40c2-97a4-56ca46df7448\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.587491 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data\") pod \"d821084c-2f3c-413c-9c98-e78f7706a75b\" (UID: \"d821084c-2f3c-413c-9c98-e78f7706a75b\") " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.588393 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.588406 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.588418 4960 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.589383 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs" (OuterVolumeSpecName: "logs") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.610150 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info" (OuterVolumeSpecName: "pod-info") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.611956 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.612769 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.613672 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs" (OuterVolumeSpecName: "certs") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.613747 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.616803 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info" (OuterVolumeSpecName: "pod-info") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.616919 4960 generic.go:334] "Generic (PLEG): container finished" podID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" containerID="7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3" exitCode=0 Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.617013 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.617330 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"21d1f29c-ee4b-4009-8017-ee502c3cfab9","Type":"ContainerDied","Data":"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.617381 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"21d1f29c-ee4b-4009-8017-ee502c3cfab9","Type":"ContainerDied","Data":"7026f6c36da5fc3be046075c10236ac5282cec17eabcea341248872260154650"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.617400 4960 scope.go:117] "RemoveContainer" containerID="7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.623294 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6" (OuterVolumeSpecName: "kube-api-access-jrpz6") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "kube-api-access-jrpz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.623693 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.628335 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.630408 4960 generic.go:334] "Generic (PLEG): container finished" podID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerID="6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac" exitCode=0 Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.630491 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerDied","Data":"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.630520 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"d821084c-2f3c-413c-9c98-e78f7706a75b","Type":"ContainerDied","Data":"df74ff1da0635fa806a304ebb7fc894f4c0405c7f4b3cedf87fbe86b729cd877"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.630581 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.642691 4960 generic.go:334] "Generic (PLEG): container finished" podID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerID="85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a" exitCode=0 Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.642777 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerDied","Data":"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.642810 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"87ca6976-ad61-40c2-97a4-56ca46df7448","Type":"ContainerDied","Data":"bb8ea855b9728ce71426fad40c31513b5bd8e1c5e75002956a4f5345eb236daf"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.642882 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.644731 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.644898 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts" (OuterVolumeSpecName: "scripts") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.652606 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8" (OuterVolumeSpecName: "kube-api-access-xssk8") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "kube-api-access-xssk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.655509 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.669008 4960 generic.go:334] "Generic (PLEG): container finished" podID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerID="3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4" exitCode=0 Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.669140 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.669227 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.669280 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerDied","Data":"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.669308 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c2b84514-def8-482d-a4a3-8b71808fcdcc","Type":"ContainerDied","Data":"41f45c327a130a59df07f3fce5873893aae2aaca7e32e3d0a7c14822ea43c4be"} Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702638 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702674 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xssk8\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-kube-api-access-xssk8\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702683 4960 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/87ca6976-ad61-40c2-97a4-56ca46df7448-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702695 4960 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702704 4960 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c2b84514-def8-482d-a4a3-8b71808fcdcc-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702712 4960 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d821084c-2f3c-413c-9c98-e78f7706a75b-logs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702722 4960 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/87ca6976-ad61-40c2-97a4-56ca46df7448-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702732 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702739 4960 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-scripts\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702750 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702760 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrpz6\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-kube-api-access-jrpz6\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702771 4960 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c2b84514-def8-482d-a4a3-8b71808fcdcc-pod-info\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702779 4960 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702788 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.702803 4960 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/projected/d821084c-2f3c-413c-9c98-e78f7706a75b-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.712416 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.736017 4960 scope.go:117] "RemoveContainer" containerID="7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.737662 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.740895 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5" (OuterVolumeSpecName: "kube-api-access-2shq5") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "kube-api-access-2shq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.742841 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3\": container with ID starting with 7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3 not found: ID does not exist" containerID="7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.744543 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3"} err="failed to get container status \"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3\": rpc error: code = NotFound desc = could not find container \"7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3\": container with ID starting with 7154098790fc7f5271b59adc0dde9c4c47f406deeae3d7781e92b48d918de1d3 not found: ID does not exist" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.744701 4960 scope.go:117] "RemoveContainer" containerID="6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.749885 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.750736 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" containerName="cloudkitty-proc" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.750834 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" containerName="cloudkitty-proc" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.750938 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751021 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751103 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c330c49-1896-4917-a0cf-4d5d5f59f25c" containerName="cloudkitty-storageinit" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751189 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c330c49-1896-4917-a0cf-4d5d5f59f25c" containerName="cloudkitty-storageinit" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751287 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751366 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751476 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api-log" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751559 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api-log" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751652 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="setup-container" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751732 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="setup-container" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751819 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.751894 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.751971 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="setup-container" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.752046 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="setup-container" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.756701 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.756817 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c330c49-1896-4917-a0cf-4d5d5f59f25c" containerName="cloudkitty-storageinit" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.756916 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.757019 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" containerName="cloudkitty-api-log" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.757126 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" containerName="rabbitmq" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.757224 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" containerName="cloudkitty-proc" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.752539 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb" (OuterVolumeSpecName: "persistence") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.758397 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.759460 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.765575 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-proc-config-data" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.783491 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c" (OuterVolumeSpecName: "persistence") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "pvc-af651932-7f63-4349-ac4a-890b6134422c". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.805309 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") on node \"crc\" " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.805353 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") on node \"crc\" " Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.805411 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2shq5\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-kube-api-access-2shq5\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.818925 4960 scope.go:117] "RemoveContainer" containerID="29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.845890 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data" (OuterVolumeSpecName: "config-data") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.849339 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data" (OuterVolumeSpecName: "config-data") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.858666 4960 scope.go:117] "RemoveContainer" containerID="6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.862241 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac\": container with ID starting with 6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac not found: ID does not exist" containerID="6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.862292 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac"} err="failed to get container status \"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac\": rpc error: code = NotFound desc = could not find container \"6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac\": container with ID starting with 6dd4ccd622e0991f46c61099b196b08497b81b6c570959b342d8c4c94d6dcdac not found: ID does not exist" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.862323 4960 scope.go:117] "RemoveContainer" containerID="29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.862663 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.863158 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-af651932-7f63-4349-ac4a-890b6134422c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c") on node "crc" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.866844 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a\": container with ID starting with 29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a not found: ID does not exist" containerID="29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.866898 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a"} err="failed to get container status \"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a\": rpc error: code = NotFound desc = could not find container \"29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a\": container with ID starting with 29117cee40b0b5d53f05d96b399cfeefa6b65a3666324742b3683165d220a17a not found: ID does not exist" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.866928 4960 scope.go:117] "RemoveContainer" containerID="85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.873762 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data" (OuterVolumeSpecName: "config-data") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.874833 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.898162 4960 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.900635 4960 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb") on node "crc" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.908297 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.908402 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-certs\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.911792 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.911988 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsj45\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-kube-api-access-zsj45\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912041 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-scripts\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912071 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912467 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912487 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912513 4960 reconciler_common.go:293] "Volume detached for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912523 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912594 4960 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.912616 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.913137 4960 scope.go:117] "RemoveContainer" containerID="d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.918423 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf" (OuterVolumeSpecName: "server-conf") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.923187 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf" (OuterVolumeSpecName: "server-conf") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.924612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.945325 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d821084c-2f3c-413c-9c98-e78f7706a75b" (UID: "d821084c-2f3c-413c-9c98-e78f7706a75b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.957246 4960 scope.go:117] "RemoveContainer" containerID="85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.959398 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a\": container with ID starting with 85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a not found: ID does not exist" containerID="85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.959440 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a"} err="failed to get container status \"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a\": rpc error: code = NotFound desc = could not find container \"85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a\": container with ID starting with 85a7586ac747bdfa35b86a4e1b679ae78d3512b239576873e82dea88122c6f3a not found: ID does not exist" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.959476 4960 scope.go:117] "RemoveContainer" containerID="d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5" Dec 01 20:05:47 crc kubenswrapper[4960]: E1201 20:05:47.959727 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5\": container with ID starting with d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5 not found: ID does not exist" containerID="d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.959749 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5"} err="failed to get container status \"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5\": rpc error: code = NotFound desc = could not find container \"d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5\": container with ID starting with d0b7d05ba604c8d0597a79ed1ac5b18ed80aa081fee998a67aba4fb40040b3a5 not found: ID does not exist" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.959767 4960 scope.go:117] "RemoveContainer" containerID="3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4" Dec 01 20:05:47 crc kubenswrapper[4960]: I1201 20:05:47.974459 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "87ca6976-ad61-40c2-97a4-56ca46df7448" (UID: "87ca6976-ad61-40c2-97a4-56ca46df7448"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.005318 4960 scope.go:117] "RemoveContainer" containerID="091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014160 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-certs\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014221 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014287 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsj45\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-kube-api-access-zsj45\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014331 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-scripts\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014360 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014422 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014486 4960 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c2b84514-def8-482d-a4a3-8b71808fcdcc-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014498 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/87ca6976-ad61-40c2-97a4-56ca46df7448-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014508 4960 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014518 4960 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/87ca6976-ad61-40c2-97a4-56ca46df7448-server-conf\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.014530 4960 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d821084c-2f3c-413c-9c98-e78f7706a75b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.021998 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c2b84514-def8-482d-a4a3-8b71808fcdcc" (UID: "c2b84514-def8-482d-a4a3-8b71808fcdcc"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.022445 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.022562 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-config-data-custom\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.022572 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-combined-ca-bundle\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.023480 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-scripts\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.025412 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-certs\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.036175 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsj45\" (UniqueName: \"kubernetes.io/projected/98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a-kube-api-access-zsj45\") pod \"cloudkitty-proc-0\" (UID: \"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a\") " pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.043872 4960 scope.go:117] "RemoveContainer" containerID="3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4" Dec 01 20:05:48 crc kubenswrapper[4960]: E1201 20:05:48.045845 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4\": container with ID starting with 3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4 not found: ID does not exist" containerID="3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.045890 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4"} err="failed to get container status \"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4\": rpc error: code = NotFound desc = could not find container \"3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4\": container with ID starting with 3a40a5244fae91866fd89b6db260fad6fb21edf589a7e66a3612f547c3da9ea4 not found: ID does not exist" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.045916 4960 scope.go:117] "RemoveContainer" containerID="091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577" Dec 01 20:05:48 crc kubenswrapper[4960]: E1201 20:05:48.046301 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577\": container with ID starting with 091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577 not found: ID does not exist" containerID="091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.046322 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577"} err="failed to get container status \"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577\": rpc error: code = NotFound desc = could not find container \"091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577\": container with ID starting with 091ffea26c40aa8f76480aac3e462100d26f85da0581c2ac0c9cdfe522307577 not found: ID does not exist" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.110988 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-proc-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.116439 4960 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c2b84514-def8-482d-a4a3-8b71808fcdcc-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.284330 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.307896 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.368434 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.411541 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.428757 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-public-svc" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.430608 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cloudkitty-internal-svc" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.468113 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cloudkitty-api-config-data" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.476292 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.572994 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.573272 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.573307 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.573415 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.573528 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.573566 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.577509 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.582501 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-logs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.582600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.582650 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6zmw\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-kube-api-access-h6zmw\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.596184 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.617009 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.649911 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.666881 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.678229 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.678241 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683306 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-57t9l" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683344 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683464 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683306 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683664 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683830 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.683895 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685066 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685170 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685188 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685220 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685244 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685263 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685296 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-logs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685315 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.685335 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6zmw\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-kube-api-access-h6zmw\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.689578 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.689741 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.689896 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.690000 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-l27tk" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.691907 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-scripts\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.691961 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.692163 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-public-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.692652 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-logs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.695285 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-combined-ca-bundle\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.695467 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data-custom\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.699256 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.699828 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.699983 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.700084 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.708337 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-config-data\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.711868 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.714260 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.717414 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-internal-tls-certs\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.723630 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6zmw\" (UniqueName: \"kubernetes.io/projected/9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf-kube-api-access-h6zmw\") pod \"cloudkitty-api-0\" (UID: \"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf\") " pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788683 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788752 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788773 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788796 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-config-data\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788825 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788851 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788871 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788899 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8rlm\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-kube-api-access-k8rlm\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788939 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788961 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.788979 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789023 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtf8q\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-kube-api-access-gtf8q\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789050 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789081 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789115 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789156 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789201 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789253 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789293 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789316 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789356 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.789382 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.891931 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8rlm\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-kube-api-access-k8rlm\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892018 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892044 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892071 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892139 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtf8q\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-kube-api-access-gtf8q\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892169 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892202 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892255 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892317 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892365 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892389 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892415 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892490 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892512 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892576 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892631 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892658 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892691 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-config-data\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892735 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.892803 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.893263 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.893264 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.895885 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-config-data\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.896078 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.896725 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.896968 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.898806 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.903778 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.903819 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d5d5199b574f54ea5db19b4cd834168e52f09a99fc0718b4c879e6f4546f0a6d/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.904464 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.906899 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.911324 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.911691 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.915243 4960 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.915451 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0004aa7191d203af0e3c5da42d773fcf606e2a73631a908716312932b18ed0b8/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.915729 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.915778 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.919932 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.921535 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.923620 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.923654 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.923979 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cloudkitty-api-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.924791 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.926884 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.929080 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.934474 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.940410 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtf8q\" (UniqueName: \"kubernetes.io/projected/05a2ef84-5a97-4844-b037-2fcdf7fba8e4-kube-api-access-gtf8q\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:48 crc kubenswrapper[4960]: I1201 20:05:48.956722 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.005825 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8rlm\" (UniqueName: \"kubernetes.io/projected/2f78cdc6-c6ed-4c26-9e66-f42bc510098d-kube-api-access-k8rlm\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.049101 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-eb4f2edf-b9ca-4b69-bfe4-7126adfd24bb\") pod \"rabbitmq-cell1-server-0\" (UID: \"05a2ef84-5a97-4844-b037-2fcdf7fba8e4\") " pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.052039 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-proc-0"] Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.065779 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-af651932-7f63-4349-ac4a-890b6134422c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-af651932-7f63-4349-ac4a-890b6134422c\") pod \"rabbitmq-server-0\" (UID: \"2f78cdc6-c6ed-4c26-9e66-f42bc510098d\") " pod="openstack/rabbitmq-server-0" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107172 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107251 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107313 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107430 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107456 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.107496 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7j6t\" (UniqueName: \"kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.112709 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.117712 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214056 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7j6t\" (UniqueName: \"kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214205 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214237 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214374 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.214392 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.216023 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.217479 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.218186 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.218988 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.219518 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.219563 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.247178 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7j6t\" (UniqueName: \"kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t\") pod \"dnsmasq-dns-dbb88bf8c-v848r\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.330005 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.373537 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d1f29c-ee4b-4009-8017-ee502c3cfab9" path="/var/lib/kubelet/pods/21d1f29c-ee4b-4009-8017-ee502c3cfab9/volumes" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.374542 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ca6976-ad61-40c2-97a4-56ca46df7448" path="/var/lib/kubelet/pods/87ca6976-ad61-40c2-97a4-56ca46df7448/volumes" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.375316 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b84514-def8-482d-a4a3-8b71808fcdcc" path="/var/lib/kubelet/pods/c2b84514-def8-482d-a4a3-8b71808fcdcc/volumes" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.376599 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d821084c-2f3c-413c-9c98-e78f7706a75b" path="/var/lib/kubelet/pods/d821084c-2f3c-413c-9c98-e78f7706a75b/volumes" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.609995 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cloudkitty-api-0"] Dec 01 20:05:49 crc kubenswrapper[4960]: W1201 20:05:49.613104 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e26ecfd_4929_4047_a4a8_6ce5b5d9a3bf.slice/crio-b59c74b46e46b6cfcccd604318d419e90d6c990d4eca8d1787df7373314e29b3 WatchSource:0}: Error finding container b59c74b46e46b6cfcccd604318d419e90d6c990d4eca8d1787df7373314e29b3: Status 404 returned error can't find the container with id b59c74b46e46b6cfcccd604318d419e90d6c990d4eca8d1787df7373314e29b3 Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.732568 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf","Type":"ContainerStarted","Data":"b59c74b46e46b6cfcccd604318d419e90d6c990d4eca8d1787df7373314e29b3"} Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.740364 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a","Type":"ContainerStarted","Data":"7484bdbe022f4e8a59ce41324742e5a794d7f55fbac5c498dea44e93d5bda77a"} Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.740449 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-proc-0" event={"ID":"98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a","Type":"ContainerStarted","Data":"8627fea9c02d6777ca421166148b24a9048dead26485052f70946446592afead"} Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.763480 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-proc-0" podStartSLOduration=2.507314076 podStartE2EDuration="2.763461525s" podCreationTimestamp="2025-12-01 20:05:47 +0000 UTC" firstStartedPulling="2025-12-01 20:05:49.02242898 +0000 UTC m=+1584.309920649" lastFinishedPulling="2025-12-01 20:05:49.278576429 +0000 UTC m=+1584.566068098" observedRunningTime="2025-12-01 20:05:49.755869309 +0000 UTC m=+1585.043360998" watchObservedRunningTime="2025-12-01 20:05:49.763461525 +0000 UTC m=+1585.050953194" Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.853637 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 01 20:05:49 crc kubenswrapper[4960]: W1201 20:05:49.876739 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05a2ef84_5a97_4844_b037_2fcdf7fba8e4.slice/crio-c1aeff67309a3c8312786142e9c339045c8f7670c74fc0abad101ba275f2706d WatchSource:0}: Error finding container c1aeff67309a3c8312786142e9c339045c8f7670c74fc0abad101ba275f2706d: Status 404 returned error can't find the container with id c1aeff67309a3c8312786142e9c339045c8f7670c74fc0abad101ba275f2706d Dec 01 20:05:49 crc kubenswrapper[4960]: I1201 20:05:49.888672 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.029203 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.751453 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05a2ef84-5a97-4844-b037-2fcdf7fba8e4","Type":"ContainerStarted","Data":"c1aeff67309a3c8312786142e9c339045c8f7670c74fc0abad101ba275f2706d"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.755516 4960 generic.go:334] "Generic (PLEG): container finished" podID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerID="00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d" exitCode=0 Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.755576 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" event={"ID":"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1","Type":"ContainerDied","Data":"00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.755603 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" event={"ID":"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1","Type":"ContainerStarted","Data":"3435988ceaac2f58055ddba042327ae4601a6bdc4b9111dc8241a15d893441e7"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.762483 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf","Type":"ContainerStarted","Data":"1aee97180810806ab0834960a87c87322b3d55ea7eac8ad5a79e2ee6714482ba"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.762518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cloudkitty-api-0" event={"ID":"9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf","Type":"ContainerStarted","Data":"6dd9639483bec637ea0764c1c180f1a27838a6b4442c1f093d6afbdca7aaf72d"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.762548 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cloudkitty-api-0" Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.772398 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2f78cdc6-c6ed-4c26-9e66-f42bc510098d","Type":"ContainerStarted","Data":"4753b767aea0f870f0733480624be51880def944a7216873c88c1f7002105b34"} Dec 01 20:05:50 crc kubenswrapper[4960]: I1201 20:05:50.811480 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cloudkitty-api-0" podStartSLOduration=2.811455504 podStartE2EDuration="2.811455504s" podCreationTimestamp="2025-12-01 20:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:05:50.803071432 +0000 UTC m=+1586.090563101" watchObservedRunningTime="2025-12-01 20:05:50.811455504 +0000 UTC m=+1586.098947173" Dec 01 20:05:51 crc kubenswrapper[4960]: I1201 20:05:51.786236 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" event={"ID":"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1","Type":"ContainerStarted","Data":"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b"} Dec 01 20:05:51 crc kubenswrapper[4960]: I1201 20:05:51.787142 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:51 crc kubenswrapper[4960]: I1201 20:05:51.811060 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" podStartSLOduration=3.811038034 podStartE2EDuration="3.811038034s" podCreationTimestamp="2025-12-01 20:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:05:51.804412957 +0000 UTC m=+1587.091904636" watchObservedRunningTime="2025-12-01 20:05:51.811038034 +0000 UTC m=+1587.098529703" Dec 01 20:05:52 crc kubenswrapper[4960]: I1201 20:05:52.802417 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05a2ef84-5a97-4844-b037-2fcdf7fba8e4","Type":"ContainerStarted","Data":"b9d5772178093d9c35e62151d8c06f734090b56e58a7bbcfcf16b76e9038f76d"} Dec 01 20:05:52 crc kubenswrapper[4960]: I1201 20:05:52.805353 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2f78cdc6-c6ed-4c26-9e66-f42bc510098d","Type":"ContainerStarted","Data":"1b8b15fd1d01758a13046e075efc05bfe6a30eacac1b0057a5a7183a7f17d853"} Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.347724 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.435296 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.435857 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="dnsmasq-dns" containerID="cri-o://54d0f458c30b5f5153d1e15cbf8ee70dee2e36af91ccd718c542f3d7cd0d8b79" gracePeriod=10 Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.653006 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-jsbnv"] Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.664242 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.693825 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-jsbnv"] Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768311 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfbjx\" (UniqueName: \"kubernetes.io/projected/fb9bb045-541f-477a-97fc-84e8bbd8f16b-kube-api-access-zfbjx\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768383 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768425 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768446 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-svc\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768519 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768560 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.768591 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-config\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.874959 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfbjx\" (UniqueName: \"kubernetes.io/projected/fb9bb045-541f-477a-97fc-84e8bbd8f16b-kube-api-access-zfbjx\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875015 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875045 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-svc\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875213 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875246 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.875269 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-config\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.876227 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-config\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.876280 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-swift-storage-0\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.876935 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-openstack-edpm-ipam\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.876940 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-nb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.877550 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-dns-svc\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.880385 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb9bb045-541f-477a-97fc-84e8bbd8f16b-ovsdbserver-sb\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.899557 4960 generic.go:334] "Generic (PLEG): container finished" podID="17246315-db1a-4a54-ae06-2060ce778508" containerID="54d0f458c30b5f5153d1e15cbf8ee70dee2e36af91ccd718c542f3d7cd0d8b79" exitCode=0 Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.899903 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" event={"ID":"17246315-db1a-4a54-ae06-2060ce778508","Type":"ContainerDied","Data":"54d0f458c30b5f5153d1e15cbf8ee70dee2e36af91ccd718c542f3d7cd0d8b79"} Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.900765 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfbjx\" (UniqueName: \"kubernetes.io/projected/fb9bb045-541f-477a-97fc-84e8bbd8f16b-kube-api-access-zfbjx\") pod \"dnsmasq-dns-85f64749dc-jsbnv\" (UID: \"fb9bb045-541f-477a-97fc-84e8bbd8f16b\") " pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:05:59 crc kubenswrapper[4960]: I1201 20:05:59.987213 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.193581 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.293876 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.293948 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.294156 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.294222 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.294286 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v656\" (UniqueName: \"kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.294895 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0\") pod \"17246315-db1a-4a54-ae06-2060ce778508\" (UID: \"17246315-db1a-4a54-ae06-2060ce778508\") " Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.301640 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656" (OuterVolumeSpecName: "kube-api-access-5v656") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "kube-api-access-5v656". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.352264 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.361526 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.363958 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.374361 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config" (OuterVolumeSpecName: "config") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.375457 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17246315-db1a-4a54-ae06-2060ce778508" (UID: "17246315-db1a-4a54-ae06-2060ce778508"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397820 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397855 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397865 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397874 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397882 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v656\" (UniqueName: \"kubernetes.io/projected/17246315-db1a-4a54-ae06-2060ce778508-kube-api-access-5v656\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.397894 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17246315-db1a-4a54-ae06-2060ce778508-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.521985 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85f64749dc-jsbnv"] Dec 01 20:06:00 crc kubenswrapper[4960]: W1201 20:06:00.523700 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb9bb045_541f_477a_97fc_84e8bbd8f16b.slice/crio-a372bc583d46844b4897158665e253c2fe503f0a3815e861f453a0e18a0d0f59 WatchSource:0}: Error finding container a372bc583d46844b4897158665e253c2fe503f0a3815e861f453a0e18a0d0f59: Status 404 returned error can't find the container with id a372bc583d46844b4897158665e253c2fe503f0a3815e861f453a0e18a0d0f59 Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.910099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" event={"ID":"fb9bb045-541f-477a-97fc-84e8bbd8f16b","Type":"ContainerStarted","Data":"a372bc583d46844b4897158665e253c2fe503f0a3815e861f453a0e18a0d0f59"} Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.912626 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" event={"ID":"17246315-db1a-4a54-ae06-2060ce778508","Type":"ContainerDied","Data":"c84aff5d2a2b5b510dd4508809b77b6b6bb4679335bda84cd4b4c28a0c21a63d"} Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.912669 4960 scope.go:117] "RemoveContainer" containerID="54d0f458c30b5f5153d1e15cbf8ee70dee2e36af91ccd718c542f3d7cd0d8b79" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.912696 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fd9b586ff-rh624" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.943426 4960 scope.go:117] "RemoveContainer" containerID="1ed6d867f843e39a4d3a63e8ee631e90d76949c4ce088870071d0acbe9bf9362" Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.966899 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:06:00 crc kubenswrapper[4960]: I1201 20:06:00.979658 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fd9b586ff-rh624"] Dec 01 20:06:01 crc kubenswrapper[4960]: I1201 20:06:01.344746 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17246315-db1a-4a54-ae06-2060ce778508" path="/var/lib/kubelet/pods/17246315-db1a-4a54-ae06-2060ce778508/volumes" Dec 01 20:06:01 crc kubenswrapper[4960]: I1201 20:06:01.931359 4960 generic.go:334] "Generic (PLEG): container finished" podID="fb9bb045-541f-477a-97fc-84e8bbd8f16b" containerID="18c3770d5f1fa94bb2099080c58a386a36ed109bed26a90526095d608cc84449" exitCode=0 Dec 01 20:06:01 crc kubenswrapper[4960]: I1201 20:06:01.931438 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" event={"ID":"fb9bb045-541f-477a-97fc-84e8bbd8f16b","Type":"ContainerDied","Data":"18c3770d5f1fa94bb2099080c58a386a36ed109bed26a90526095d608cc84449"} Dec 01 20:06:02 crc kubenswrapper[4960]: I1201 20:06:02.945296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" event={"ID":"fb9bb045-541f-477a-97fc-84e8bbd8f16b","Type":"ContainerStarted","Data":"aab8888802e0a84fc3974ad396a4cbc102214f346b1af7a6979f3ace695d68e7"} Dec 01 20:06:02 crc kubenswrapper[4960]: I1201 20:06:02.945653 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:06:02 crc kubenswrapper[4960]: I1201 20:06:02.968106 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" podStartSLOduration=3.96808543 podStartE2EDuration="3.96808543s" podCreationTimestamp="2025-12-01 20:05:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:06:02.961924998 +0000 UTC m=+1598.249416677" watchObservedRunningTime="2025-12-01 20:06:02.96808543 +0000 UTC m=+1598.255577099" Dec 01 20:06:06 crc kubenswrapper[4960]: I1201 20:06:06.860992 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 01 20:06:09 crc kubenswrapper[4960]: I1201 20:06:09.991332 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85f64749dc-jsbnv" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.056979 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.057234 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="dnsmasq-dns" containerID="cri-o://6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b" gracePeriod=10 Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.626163 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.725843 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726098 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7j6t\" (UniqueName: \"kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726196 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726365 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726456 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726639 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.726697 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0\") pod \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\" (UID: \"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1\") " Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.732300 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t" (OuterVolumeSpecName: "kube-api-access-j7j6t") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "kube-api-access-j7j6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.801987 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.806004 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.806823 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.812821 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.815371 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.818567 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config" (OuterVolumeSpecName: "config") pod "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" (UID: "637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830288 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830391 4960 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830478 4960 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830630 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7j6t\" (UniqueName: \"kubernetes.io/projected/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-kube-api-access-j7j6t\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830688 4960 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830746 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:10 crc kubenswrapper[4960]: I1201 20:06:10.830798 4960 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.060574 4960 generic.go:334] "Generic (PLEG): container finished" podID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerID="6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b" exitCode=0 Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.060618 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" event={"ID":"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1","Type":"ContainerDied","Data":"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b"} Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.060647 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" event={"ID":"637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1","Type":"ContainerDied","Data":"3435988ceaac2f58055ddba042327ae4601a6bdc4b9111dc8241a15d893441e7"} Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.060665 4960 scope.go:117] "RemoveContainer" containerID="6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.060837 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dbb88bf8c-v848r" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.114612 4960 scope.go:117] "RemoveContainer" containerID="00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.128214 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.138431 4960 scope.go:117] "RemoveContainer" containerID="6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b" Dec 01 20:06:11 crc kubenswrapper[4960]: E1201 20:06:11.138942 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b\": container with ID starting with 6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b not found: ID does not exist" containerID="6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.138992 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b"} err="failed to get container status \"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b\": rpc error: code = NotFound desc = could not find container \"6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b\": container with ID starting with 6b095e7c64737862480c5900c09d9d74ab2c4ba67e4e8847706822e1899b656b not found: ID does not exist" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.139021 4960 scope.go:117] "RemoveContainer" containerID="00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d" Dec 01 20:06:11 crc kubenswrapper[4960]: E1201 20:06:11.139329 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d\": container with ID starting with 00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d not found: ID does not exist" containerID="00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.139365 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d"} err="failed to get container status \"00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d\": rpc error: code = NotFound desc = could not find container \"00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d\": container with ID starting with 00831400f78b0bcc2565ce35852a6cace66ba80bcf4035c24d209158aa11a22d not found: ID does not exist" Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.139422 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dbb88bf8c-v848r"] Dec 01 20:06:11 crc kubenswrapper[4960]: I1201 20:06:11.337963 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" path="/var/lib/kubelet/pods/637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1/volumes" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.494539 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n"] Dec 01 20:06:23 crc kubenswrapper[4960]: E1201 20:06:23.495758 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.495779 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: E1201 20:06:23.495796 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.495805 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: E1201 20:06:23.495826 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="init" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.495834 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="init" Dec 01 20:06:23 crc kubenswrapper[4960]: E1201 20:06:23.495849 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="init" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.495856 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="init" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.496153 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="637acb2e-09e3-4d40-a8cb-fb8aa91c4ca1" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.496220 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="17246315-db1a-4a54-ae06-2060ce778508" containerName="dnsmasq-dns" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.497213 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.502697 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.502995 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.503159 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.503440 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.508622 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n"] Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.590177 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqcwz\" (UniqueName: \"kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.590372 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.590437 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.590481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.692141 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqcwz\" (UniqueName: \"kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.692529 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.692641 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.693328 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.699896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.699925 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.700239 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.707818 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqcwz\" (UniqueName: \"kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:23 crc kubenswrapper[4960]: I1201 20:06:23.824343 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:24 crc kubenswrapper[4960]: I1201 20:06:24.203353 4960 generic.go:334] "Generic (PLEG): container finished" podID="2f78cdc6-c6ed-4c26-9e66-f42bc510098d" containerID="1b8b15fd1d01758a13046e075efc05bfe6a30eacac1b0057a5a7183a7f17d853" exitCode=0 Dec 01 20:06:24 crc kubenswrapper[4960]: I1201 20:06:24.203428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2f78cdc6-c6ed-4c26-9e66-f42bc510098d","Type":"ContainerDied","Data":"1b8b15fd1d01758a13046e075efc05bfe6a30eacac1b0057a5a7183a7f17d853"} Dec 01 20:06:24 crc kubenswrapper[4960]: I1201 20:06:24.475474 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n"] Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.220468 4960 generic.go:334] "Generic (PLEG): container finished" podID="05a2ef84-5a97-4844-b037-2fcdf7fba8e4" containerID="b9d5772178093d9c35e62151d8c06f734090b56e58a7bbcfcf16b76e9038f76d" exitCode=0 Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.220529 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05a2ef84-5a97-4844-b037-2fcdf7fba8e4","Type":"ContainerDied","Data":"b9d5772178093d9c35e62151d8c06f734090b56e58a7bbcfcf16b76e9038f76d"} Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.224954 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2f78cdc6-c6ed-4c26-9e66-f42bc510098d","Type":"ContainerStarted","Data":"072531eadb4e08c0c37d779d6018742bb75d4e1968768f669d91782e6946fd3a"} Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.225365 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.227032 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" event={"ID":"bca07824-2e7a-4578-b833-02232c52ca0c","Type":"ContainerStarted","Data":"15195e2d8f5e0de9b7bf3c6230fe0d36f476df2cbe867748e60c35e9cadd1d4c"} Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.297399 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.297374258 podStartE2EDuration="37.297374258s" podCreationTimestamp="2025-12-01 20:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:06:25.286401197 +0000 UTC m=+1620.573892876" watchObservedRunningTime="2025-12-01 20:06:25.297374258 +0000 UTC m=+1620.584865927" Dec 01 20:06:25 crc kubenswrapper[4960]: I1201 20:06:25.900609 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cloudkitty-api-0" Dec 01 20:06:26 crc kubenswrapper[4960]: I1201 20:06:26.240343 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"05a2ef84-5a97-4844-b037-2fcdf7fba8e4","Type":"ContainerStarted","Data":"0cd9003656c67620387f486874ec59939e9a587c73ed646c88e909a893feaf48"} Dec 01 20:06:26 crc kubenswrapper[4960]: I1201 20:06:26.270793 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.270775772 podStartE2EDuration="38.270775772s" podCreationTimestamp="2025-12-01 20:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:06:26.268167561 +0000 UTC m=+1621.555659310" watchObservedRunningTime="2025-12-01 20:06:26.270775772 +0000 UTC m=+1621.558267441" Dec 01 20:06:29 crc kubenswrapper[4960]: I1201 20:06:29.113817 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:06:34 crc kubenswrapper[4960]: I1201 20:06:34.367723 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:06:35 crc kubenswrapper[4960]: I1201 20:06:35.360139 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" event={"ID":"bca07824-2e7a-4578-b833-02232c52ca0c","Type":"ContainerStarted","Data":"25341a2f372b8610fbba27561cd65c5df4329f4e0c40800b8c44fca0eda2d7e5"} Dec 01 20:06:35 crc kubenswrapper[4960]: I1201 20:06:35.389306 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" podStartSLOduration=2.509021193 podStartE2EDuration="12.389285983s" podCreationTimestamp="2025-12-01 20:06:23 +0000 UTC" firstStartedPulling="2025-12-01 20:06:24.484702991 +0000 UTC m=+1619.772194670" lastFinishedPulling="2025-12-01 20:06:34.364967781 +0000 UTC m=+1629.652459460" observedRunningTime="2025-12-01 20:06:35.383456032 +0000 UTC m=+1630.670947711" watchObservedRunningTime="2025-12-01 20:06:35.389285983 +0000 UTC m=+1630.676777652" Dec 01 20:06:39 crc kubenswrapper[4960]: I1201 20:06:39.117393 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 01 20:06:39 crc kubenswrapper[4960]: I1201 20:06:39.121343 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 01 20:06:43 crc kubenswrapper[4960]: I1201 20:06:43.870490 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:06:43 crc kubenswrapper[4960]: I1201 20:06:43.871393 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:06:46 crc kubenswrapper[4960]: E1201 20:06:46.130824 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbca07824_2e7a_4578_b833_02232c52ca0c.slice/crio-25341a2f372b8610fbba27561cd65c5df4329f4e0c40800b8c44fca0eda2d7e5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbca07824_2e7a_4578_b833_02232c52ca0c.slice/crio-conmon-25341a2f372b8610fbba27561cd65c5df4329f4e0c40800b8c44fca0eda2d7e5.scope\": RecentStats: unable to find data in memory cache]" Dec 01 20:06:46 crc kubenswrapper[4960]: I1201 20:06:46.508659 4960 generic.go:334] "Generic (PLEG): container finished" podID="bca07824-2e7a-4578-b833-02232c52ca0c" containerID="25341a2f372b8610fbba27561cd65c5df4329f4e0c40800b8c44fca0eda2d7e5" exitCode=0 Dec 01 20:06:46 crc kubenswrapper[4960]: I1201 20:06:46.508718 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" event={"ID":"bca07824-2e7a-4578-b833-02232c52ca0c","Type":"ContainerDied","Data":"25341a2f372b8610fbba27561cd65c5df4329f4e0c40800b8c44fca0eda2d7e5"} Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.138853 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.263633 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key\") pod \"bca07824-2e7a-4578-b833-02232c52ca0c\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.263671 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory\") pod \"bca07824-2e7a-4578-b833-02232c52ca0c\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.263700 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle\") pod \"bca07824-2e7a-4578-b833-02232c52ca0c\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.263864 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqcwz\" (UniqueName: \"kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz\") pod \"bca07824-2e7a-4578-b833-02232c52ca0c\" (UID: \"bca07824-2e7a-4578-b833-02232c52ca0c\") " Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.270612 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bca07824-2e7a-4578-b833-02232c52ca0c" (UID: "bca07824-2e7a-4578-b833-02232c52ca0c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.270930 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz" (OuterVolumeSpecName: "kube-api-access-sqcwz") pod "bca07824-2e7a-4578-b833-02232c52ca0c" (UID: "bca07824-2e7a-4578-b833-02232c52ca0c"). InnerVolumeSpecName "kube-api-access-sqcwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.295449 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory" (OuterVolumeSpecName: "inventory") pod "bca07824-2e7a-4578-b833-02232c52ca0c" (UID: "bca07824-2e7a-4578-b833-02232c52ca0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.307587 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bca07824-2e7a-4578-b833-02232c52ca0c" (UID: "bca07824-2e7a-4578-b833-02232c52ca0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.367875 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.367921 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.367935 4960 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bca07824-2e7a-4578-b833-02232c52ca0c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.367949 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqcwz\" (UniqueName: \"kubernetes.io/projected/bca07824-2e7a-4578-b833-02232c52ca0c-kube-api-access-sqcwz\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.530763 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" event={"ID":"bca07824-2e7a-4578-b833-02232c52ca0c","Type":"ContainerDied","Data":"15195e2d8f5e0de9b7bf3c6230fe0d36f476df2cbe867748e60c35e9cadd1d4c"} Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.530811 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15195e2d8f5e0de9b7bf3c6230fe0d36f476df2cbe867748e60c35e9cadd1d4c" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.530819 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.622787 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m"] Dec 01 20:06:48 crc kubenswrapper[4960]: E1201 20:06:48.623402 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca07824-2e7a-4578-b833-02232c52ca0c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.623422 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca07824-2e7a-4578-b833-02232c52ca0c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.623704 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="bca07824-2e7a-4578-b833-02232c52ca0c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.624721 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.626829 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.627050 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.628323 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.630659 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.636633 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m"] Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.673966 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.674017 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnfnp\" (UniqueName: \"kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.674086 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.775846 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.776616 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.776764 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnfnp\" (UniqueName: \"kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.780641 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.780896 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.798523 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnfnp\" (UniqueName: \"kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ndr5m\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:48 crc kubenswrapper[4960]: I1201 20:06:48.940518 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:49 crc kubenswrapper[4960]: I1201 20:06:49.477748 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m"] Dec 01 20:06:49 crc kubenswrapper[4960]: I1201 20:06:49.543969 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" event={"ID":"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad","Type":"ContainerStarted","Data":"e12a7387c1a901a53dccd8bfe962a7eb34eab3283286eac49b514edc6fbcdf15"} Dec 01 20:06:50 crc kubenswrapper[4960]: I1201 20:06:50.558186 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" event={"ID":"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad","Type":"ContainerStarted","Data":"06268bdc496938b84ea2d7d62cba628b6473ef073112d760c8a37780f8fa7483"} Dec 01 20:06:50 crc kubenswrapper[4960]: I1201 20:06:50.576299 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" podStartSLOduration=2.117814823 podStartE2EDuration="2.576273616s" podCreationTimestamp="2025-12-01 20:06:48 +0000 UTC" firstStartedPulling="2025-12-01 20:06:49.485447443 +0000 UTC m=+1644.772939112" lastFinishedPulling="2025-12-01 20:06:49.943906236 +0000 UTC m=+1645.231397905" observedRunningTime="2025-12-01 20:06:50.574390097 +0000 UTC m=+1645.861881776" watchObservedRunningTime="2025-12-01 20:06:50.576273616 +0000 UTC m=+1645.863765285" Dec 01 20:06:53 crc kubenswrapper[4960]: I1201 20:06:53.607482 4960 generic.go:334] "Generic (PLEG): container finished" podID="03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" containerID="06268bdc496938b84ea2d7d62cba628b6473ef073112d760c8a37780f8fa7483" exitCode=0 Dec 01 20:06:53 crc kubenswrapper[4960]: I1201 20:06:53.607607 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" event={"ID":"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad","Type":"ContainerDied","Data":"06268bdc496938b84ea2d7d62cba628b6473ef073112d760c8a37780f8fa7483"} Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.095330 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.212349 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory\") pod \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.212446 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnfnp\" (UniqueName: \"kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp\") pod \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.212547 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key\") pod \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\" (UID: \"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad\") " Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.221999 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp" (OuterVolumeSpecName: "kube-api-access-jnfnp") pod "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" (UID: "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad"). InnerVolumeSpecName "kube-api-access-jnfnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.253928 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory" (OuterVolumeSpecName: "inventory") pod "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" (UID: "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.266374 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" (UID: "03306a2e-8c24-4479-bf7c-f6fbae3dd8ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.316295 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.316710 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnfnp\" (UniqueName: \"kubernetes.io/projected/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-kube-api-access-jnfnp\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.316731 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03306a2e-8c24-4479-bf7c-f6fbae3dd8ad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.633399 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" event={"ID":"03306a2e-8c24-4479-bf7c-f6fbae3dd8ad","Type":"ContainerDied","Data":"e12a7387c1a901a53dccd8bfe962a7eb34eab3283286eac49b514edc6fbcdf15"} Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.633449 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e12a7387c1a901a53dccd8bfe962a7eb34eab3283286eac49b514edc6fbcdf15" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.633512 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ndr5m" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.702410 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb"] Dec 01 20:06:55 crc kubenswrapper[4960]: E1201 20:06:55.703033 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.703063 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.703379 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="03306a2e-8c24-4479-bf7c-f6fbae3dd8ad" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.704266 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.706462 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.706563 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.706575 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.709207 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.718471 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb"] Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.827425 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmn48\" (UniqueName: \"kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.827490 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.827548 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.827708 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.929665 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmn48\" (UniqueName: \"kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.929760 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.929843 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.930266 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.936425 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.939867 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.940066 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:55 crc kubenswrapper[4960]: I1201 20:06:55.959275 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmn48\" (UniqueName: \"kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:56 crc kubenswrapper[4960]: I1201 20:06:56.029830 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:06:56 crc kubenswrapper[4960]: I1201 20:06:56.628628 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb"] Dec 01 20:06:56 crc kubenswrapper[4960]: I1201 20:06:56.646675 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" event={"ID":"098a01da-e6f9-4bc8-8249-72bbeaf949f7","Type":"ContainerStarted","Data":"74b5231f99601dc597f47ebd01aa33749fe2503fbfcb6f6c6888b74b31592338"} Dec 01 20:06:57 crc kubenswrapper[4960]: I1201 20:06:57.659132 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" event={"ID":"098a01da-e6f9-4bc8-8249-72bbeaf949f7","Type":"ContainerStarted","Data":"15c69b10a58882d5ad6f4b82607766a1ed31cd34a70759697709c85499c1a564"} Dec 01 20:06:57 crc kubenswrapper[4960]: I1201 20:06:57.681418 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" podStartSLOduration=1.962279318 podStartE2EDuration="2.681402122s" podCreationTimestamp="2025-12-01 20:06:55 +0000 UTC" firstStartedPulling="2025-12-01 20:06:56.630606306 +0000 UTC m=+1651.918097975" lastFinishedPulling="2025-12-01 20:06:57.34972911 +0000 UTC m=+1652.637220779" observedRunningTime="2025-12-01 20:06:57.675811608 +0000 UTC m=+1652.963303277" watchObservedRunningTime="2025-12-01 20:06:57.681402122 +0000 UTC m=+1652.968893791" Dec 01 20:07:11 crc kubenswrapper[4960]: I1201 20:07:11.824637 4960 scope.go:117] "RemoveContainer" containerID="8a228de05fbd4b29f1cbedc4d66e6a2ca34dccefd0c3c14d02cacaaa55c5270b" Dec 01 20:07:13 crc kubenswrapper[4960]: I1201 20:07:13.871054 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:07:13 crc kubenswrapper[4960]: I1201 20:07:13.871542 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:07:43 crc kubenswrapper[4960]: I1201 20:07:43.870972 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:07:43 crc kubenswrapper[4960]: I1201 20:07:43.871652 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:07:43 crc kubenswrapper[4960]: I1201 20:07:43.871715 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:07:43 crc kubenswrapper[4960]: I1201 20:07:43.872717 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:07:43 crc kubenswrapper[4960]: I1201 20:07:43.872786 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" gracePeriod=600 Dec 01 20:07:44 crc kubenswrapper[4960]: E1201 20:07:44.039172 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:07:44 crc kubenswrapper[4960]: I1201 20:07:44.198869 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" exitCode=0 Dec 01 20:07:44 crc kubenswrapper[4960]: I1201 20:07:44.198897 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01"} Dec 01 20:07:44 crc kubenswrapper[4960]: I1201 20:07:44.199292 4960 scope.go:117] "RemoveContainer" containerID="7373470eec9a1b95e49d63347ad6768fecc70a5caa1db6f177168a7ea254a923" Dec 01 20:07:44 crc kubenswrapper[4960]: I1201 20:07:44.200351 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:07:44 crc kubenswrapper[4960]: E1201 20:07:44.200940 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:07:59 crc kubenswrapper[4960]: I1201 20:07:59.328025 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:07:59 crc kubenswrapper[4960]: E1201 20:07:59.329005 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:08:11 crc kubenswrapper[4960]: I1201 20:08:11.902136 4960 scope.go:117] "RemoveContainer" containerID="f9b86a25e5c116940aa360b7d179569e5c387f9371bca29da93408cef36aa980" Dec 01 20:08:11 crc kubenswrapper[4960]: I1201 20:08:11.935466 4960 scope.go:117] "RemoveContainer" containerID="b67f672d0c21fd16cc2d41ae3278ab79a01c86ded2c7e9dde99c984294809c3e" Dec 01 20:08:11 crc kubenswrapper[4960]: I1201 20:08:11.992258 4960 scope.go:117] "RemoveContainer" containerID="1d8f606812105aa624867384d53a722d0010b121426d44171d1d0b7694081716" Dec 01 20:08:14 crc kubenswrapper[4960]: I1201 20:08:14.324312 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:08:14 crc kubenswrapper[4960]: E1201 20:08:14.324905 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:08:27 crc kubenswrapper[4960]: I1201 20:08:27.324723 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:08:27 crc kubenswrapper[4960]: E1201 20:08:27.325592 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:08:39 crc kubenswrapper[4960]: I1201 20:08:39.326759 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:08:39 crc kubenswrapper[4960]: E1201 20:08:39.328309 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:08:53 crc kubenswrapper[4960]: I1201 20:08:53.323920 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:08:53 crc kubenswrapper[4960]: E1201 20:08:53.325432 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:09:04 crc kubenswrapper[4960]: I1201 20:09:04.324219 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:09:04 crc kubenswrapper[4960]: E1201 20:09:04.325338 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:09:12 crc kubenswrapper[4960]: I1201 20:09:12.092034 4960 scope.go:117] "RemoveContainer" containerID="2a8938c6209003b731a5d76aa0618ea9d20d4ad73dfb361c0d2f7022bb22c571" Dec 01 20:09:12 crc kubenswrapper[4960]: I1201 20:09:12.158977 4960 scope.go:117] "RemoveContainer" containerID="0ec9baf7f65649b3b8c9c3d9d93cc816bf287dba908ac6b621e15294f8d2eaad" Dec 01 20:09:15 crc kubenswrapper[4960]: I1201 20:09:15.339783 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:09:15 crc kubenswrapper[4960]: E1201 20:09:15.340894 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:09:26 crc kubenswrapper[4960]: I1201 20:09:26.324356 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:09:26 crc kubenswrapper[4960]: E1201 20:09:26.325770 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:09:41 crc kubenswrapper[4960]: I1201 20:09:41.324857 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:09:41 crc kubenswrapper[4960]: E1201 20:09:41.325971 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:09:53 crc kubenswrapper[4960]: I1201 20:09:53.324062 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:09:53 crc kubenswrapper[4960]: E1201 20:09:53.325069 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:10:05 crc kubenswrapper[4960]: I1201 20:10:05.339637 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:10:05 crc kubenswrapper[4960]: E1201 20:10:05.340640 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:10:19 crc kubenswrapper[4960]: I1201 20:10:19.324677 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:10:19 crc kubenswrapper[4960]: E1201 20:10:19.325592 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:10:34 crc kubenswrapper[4960]: I1201 20:10:34.323917 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:10:34 crc kubenswrapper[4960]: E1201 20:10:34.325135 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.056555 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2bwq2"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.065825 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-fbe8-account-create-update-2ngk5"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.077703 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-ff26-account-create-update-xgdmp"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.087621 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5d22-account-create-update-mvkxq"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.095883 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wtzn8"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.104357 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2bwq2"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.112967 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-gdkrq"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.122220 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-fbe8-account-create-update-2ngk5"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.131008 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-gdkrq"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.139788 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-ff26-account-create-update-xgdmp"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.148267 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wtzn8"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.157026 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5d22-account-create-update-mvkxq"] Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.344540 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04fb1740-7860-4e08-9a62-13e55b59ed77" path="/var/lib/kubelet/pods/04fb1740-7860-4e08-9a62-13e55b59ed77/volumes" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.345522 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13bd72fe-2556-4e33-bf2c-f95eba6d2310" path="/var/lib/kubelet/pods/13bd72fe-2556-4e33-bf2c-f95eba6d2310/volumes" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.346582 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51c22ac6-d753-4723-8889-c839eb542bf1" path="/var/lib/kubelet/pods/51c22ac6-d753-4723-8889-c839eb542bf1/volumes" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.347558 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dec74ac-5b50-42ed-9b0e-c58f881eef9c" path="/var/lib/kubelet/pods/6dec74ac-5b50-42ed-9b0e-c58f881eef9c/volumes" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.350105 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce13c3c2-01f2-4d05-a8e8-3107617eaf29" path="/var/lib/kubelet/pods/ce13c3c2-01f2-4d05-a8e8-3107617eaf29/volumes" Dec 01 20:10:45 crc kubenswrapper[4960]: I1201 20:10:45.351411 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec2068f8-377d-454f-80f3-429d30d4ecaa" path="/var/lib/kubelet/pods/ec2068f8-377d-454f-80f3-429d30d4ecaa/volumes" Dec 01 20:10:47 crc kubenswrapper[4960]: I1201 20:10:47.324485 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:10:47 crc kubenswrapper[4960]: E1201 20:10:47.324989 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:10:53 crc kubenswrapper[4960]: I1201 20:10:53.436893 4960 generic.go:334] "Generic (PLEG): container finished" podID="098a01da-e6f9-4bc8-8249-72bbeaf949f7" containerID="15c69b10a58882d5ad6f4b82607766a1ed31cd34a70759697709c85499c1a564" exitCode=0 Dec 01 20:10:53 crc kubenswrapper[4960]: I1201 20:10:53.437296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" event={"ID":"098a01da-e6f9-4bc8-8249-72bbeaf949f7","Type":"ContainerDied","Data":"15c69b10a58882d5ad6f4b82607766a1ed31cd34a70759697709c85499c1a564"} Dec 01 20:10:54 crc kubenswrapper[4960]: I1201 20:10:54.937498 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.066573 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmn48\" (UniqueName: \"kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48\") pod \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.066647 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle\") pod \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.066756 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory\") pod \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.067841 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key\") pod \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\" (UID: \"098a01da-e6f9-4bc8-8249-72bbeaf949f7\") " Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.077395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48" (OuterVolumeSpecName: "kube-api-access-zmn48") pod "098a01da-e6f9-4bc8-8249-72bbeaf949f7" (UID: "098a01da-e6f9-4bc8-8249-72bbeaf949f7"). InnerVolumeSpecName "kube-api-access-zmn48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.083325 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "098a01da-e6f9-4bc8-8249-72bbeaf949f7" (UID: "098a01da-e6f9-4bc8-8249-72bbeaf949f7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.126245 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory" (OuterVolumeSpecName: "inventory") pod "098a01da-e6f9-4bc8-8249-72bbeaf949f7" (UID: "098a01da-e6f9-4bc8-8249-72bbeaf949f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.160723 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "098a01da-e6f9-4bc8-8249-72bbeaf949f7" (UID: "098a01da-e6f9-4bc8-8249-72bbeaf949f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.170800 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmn48\" (UniqueName: \"kubernetes.io/projected/098a01da-e6f9-4bc8-8249-72bbeaf949f7-kube-api-access-zmn48\") on node \"crc\" DevicePath \"\"" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.170835 4960 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.170845 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.170853 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/098a01da-e6f9-4bc8-8249-72bbeaf949f7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.459510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" event={"ID":"098a01da-e6f9-4bc8-8249-72bbeaf949f7","Type":"ContainerDied","Data":"74b5231f99601dc597f47ebd01aa33749fe2503fbfcb6f6c6888b74b31592338"} Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.460074 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74b5231f99601dc597f47ebd01aa33749fe2503fbfcb6f6c6888b74b31592338" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.459608 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.557851 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd"] Dec 01 20:10:55 crc kubenswrapper[4960]: E1201 20:10:55.558354 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098a01da-e6f9-4bc8-8249-72bbeaf949f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.558374 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="098a01da-e6f9-4bc8-8249-72bbeaf949f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.558593 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="098a01da-e6f9-4bc8-8249-72bbeaf949f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.559509 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.562071 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.562288 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.562422 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.563537 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.573547 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd"] Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.681045 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgdhk\" (UniqueName: \"kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.681182 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.681275 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.782895 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.782991 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgdhk\" (UniqueName: \"kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.783079 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.787575 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.788062 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.801290 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgdhk\" (UniqueName: \"kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:55 crc kubenswrapper[4960]: I1201 20:10:55.876435 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:10:56 crc kubenswrapper[4960]: I1201 20:10:56.433375 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd"] Dec 01 20:10:56 crc kubenswrapper[4960]: W1201 20:10:56.437755 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod476f3fb9_9214_4448_9ff4_a5489883199b.slice/crio-6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9 WatchSource:0}: Error finding container 6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9: Status 404 returned error can't find the container with id 6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9 Dec 01 20:10:56 crc kubenswrapper[4960]: I1201 20:10:56.440304 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:10:56 crc kubenswrapper[4960]: I1201 20:10:56.468010 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" event={"ID":"476f3fb9-9214-4448-9ff4-a5489883199b","Type":"ContainerStarted","Data":"6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9"} Dec 01 20:10:58 crc kubenswrapper[4960]: I1201 20:10:58.496496 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" event={"ID":"476f3fb9-9214-4448-9ff4-a5489883199b","Type":"ContainerStarted","Data":"eeacfd487a6cc2a4509193a76da737ce98a655e2c974ffe3a74e576191594a17"} Dec 01 20:10:58 crc kubenswrapper[4960]: I1201 20:10:58.518679 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" podStartSLOduration=2.330627345 podStartE2EDuration="3.518658795s" podCreationTimestamp="2025-12-01 20:10:55 +0000 UTC" firstStartedPulling="2025-12-01 20:10:56.440026325 +0000 UTC m=+1891.727518004" lastFinishedPulling="2025-12-01 20:10:57.628057785 +0000 UTC m=+1892.915549454" observedRunningTime="2025-12-01 20:10:58.515886278 +0000 UTC m=+1893.803377967" watchObservedRunningTime="2025-12-01 20:10:58.518658795 +0000 UTC m=+1893.806150464" Dec 01 20:11:01 crc kubenswrapper[4960]: I1201 20:11:01.324339 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:11:01 crc kubenswrapper[4960]: E1201 20:11:01.325160 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.304463 4960 scope.go:117] "RemoveContainer" containerID="dc30df6ecc7121cad23b772db529e2a007209ede885b3c396bdf0c333936f0b6" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.335606 4960 scope.go:117] "RemoveContainer" containerID="d04fed7483633c06890d66872983a3cf3122ed60a9748ac1de8cfcc692cccbbc" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.407773 4960 scope.go:117] "RemoveContainer" containerID="ff28f8e07e09ab9cd1123e1820ca7f55f0fdb1ecc1dd405fa046a7183daa23d1" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.447252 4960 scope.go:117] "RemoveContainer" containerID="fff22dcc570d6e3492f343e709800b403f67337ea4d45d409d6c5bc1f472086e" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.504942 4960 scope.go:117] "RemoveContainer" containerID="5fa51d98ab83a1a11058214ec066e19f95031fbc4be912242a13d584f741cfff" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.535893 4960 scope.go:117] "RemoveContainer" containerID="7a9c7dd926d087fc72804d3ce9caf7d55a31fbeb6e47935be41f7689ff48ab6f" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.573788 4960 scope.go:117] "RemoveContainer" containerID="2c3189f656ce75428367a615a03c7efc9c7e8ae0409bb7d370ac2a80de90ca2a" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.605361 4960 scope.go:117] "RemoveContainer" containerID="46ccfc853f030042a7463e7ac6d328af94ad03c4a4a753919a7ed53a294dbe62" Dec 01 20:11:12 crc kubenswrapper[4960]: I1201 20:11:12.629456 4960 scope.go:117] "RemoveContainer" containerID="f55900b7072a38b367473834cf0325067c71eadd1305de7e1f6b8fdd4834a91b" Dec 01 20:11:13 crc kubenswrapper[4960]: I1201 20:11:13.057078 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-r6j4l"] Dec 01 20:11:13 crc kubenswrapper[4960]: I1201 20:11:13.073953 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-r6j4l"] Dec 01 20:11:13 crc kubenswrapper[4960]: I1201 20:11:13.342278 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec" path="/var/lib/kubelet/pods/ad33b2c4-4c4f-41b0-9e3f-6455bf3dfeec/volumes" Dec 01 20:11:16 crc kubenswrapper[4960]: I1201 20:11:16.324712 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:11:16 crc kubenswrapper[4960]: E1201 20:11:16.325569 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.032741 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qb9hb"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.046859 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qb9hb"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.057310 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8eff-account-create-update-2p9kt"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.066216 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-9a0f-account-create-update-brw2l"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.076232 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-frxcr"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.085271 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8eff-account-create-update-2p9kt"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.093421 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-9a0f-account-create-update-brw2l"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.101455 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-create-fbxww"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.109494 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-z7sq7"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.117584 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-frxcr"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.125503 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-z7sq7"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.133691 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-create-fbxww"] Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.337844 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="923a3511-6d9b-4143-867b-552769cdbd94" path="/var/lib/kubelet/pods/923a3511-6d9b-4143-867b-552769cdbd94/volumes" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.338843 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b513d8e1-b2fa-476d-b925-12ce32db7390" path="/var/lib/kubelet/pods/b513d8e1-b2fa-476d-b925-12ce32db7390/volumes" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.339667 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd132ae6-3b94-42f6-804a-29e58ba2d671" path="/var/lib/kubelet/pods/bd132ae6-3b94-42f6-804a-29e58ba2d671/volumes" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.340441 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ce664d-4e8c-479a-95af-9f38cdeb98eb" path="/var/lib/kubelet/pods/c2ce664d-4e8c-479a-95af-9f38cdeb98eb/volumes" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.341821 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ead81fb5-9e98-4b2a-8e02-8d9b82bef13b" path="/var/lib/kubelet/pods/ead81fb5-9e98-4b2a-8e02-8d9b82bef13b/volumes" Dec 01 20:11:19 crc kubenswrapper[4960]: I1201 20:11:19.342572 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe342ccd-a32e-42ba-9018-c4fe438c9d8c" path="/var/lib/kubelet/pods/fe342ccd-a32e-42ba-9018-c4fe438c9d8c/volumes" Dec 01 20:11:20 crc kubenswrapper[4960]: I1201 20:11:20.028755 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-87eb-account-create-update-xw6fw"] Dec 01 20:11:20 crc kubenswrapper[4960]: I1201 20:11:20.038845 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-cc43-account-create-update-kkd2v"] Dec 01 20:11:20 crc kubenswrapper[4960]: I1201 20:11:20.047943 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-cc43-account-create-update-kkd2v"] Dec 01 20:11:20 crc kubenswrapper[4960]: I1201 20:11:20.056831 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-87eb-account-create-update-xw6fw"] Dec 01 20:11:21 crc kubenswrapper[4960]: I1201 20:11:21.338654 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0190c0ae-497c-404f-b7f0-a57250e0339e" path="/var/lib/kubelet/pods/0190c0ae-497c-404f-b7f0-a57250e0339e/volumes" Dec 01 20:11:21 crc kubenswrapper[4960]: I1201 20:11:21.339229 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b811ebc-3885-4b30-8e97-373aaaff228e" path="/var/lib/kubelet/pods/8b811ebc-3885-4b30-8e97-373aaaff228e/volumes" Dec 01 20:11:29 crc kubenswrapper[4960]: I1201 20:11:29.036069 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-fl9dv"] Dec 01 20:11:29 crc kubenswrapper[4960]: I1201 20:11:29.047211 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-fl9dv"] Dec 01 20:11:29 crc kubenswrapper[4960]: I1201 20:11:29.335476 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="948143c4-753e-4129-9712-825a69981406" path="/var/lib/kubelet/pods/948143c4-753e-4129-9712-825a69981406/volumes" Dec 01 20:11:31 crc kubenswrapper[4960]: I1201 20:11:31.325744 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:11:31 crc kubenswrapper[4960]: E1201 20:11:31.326439 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:11:44 crc kubenswrapper[4960]: I1201 20:11:44.324665 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:11:44 crc kubenswrapper[4960]: E1201 20:11:44.325486 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:11:58 crc kubenswrapper[4960]: I1201 20:11:58.324334 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:11:58 crc kubenswrapper[4960]: E1201 20:11:58.325209 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:12:00 crc kubenswrapper[4960]: I1201 20:12:00.072994 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-26xd2"] Dec 01 20:12:00 crc kubenswrapper[4960]: I1201 20:12:00.085510 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-26xd2"] Dec 01 20:12:01 crc kubenswrapper[4960]: I1201 20:12:01.342849 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e3f17e-3310-4ef3-8de1-637bf905a126" path="/var/lib/kubelet/pods/69e3f17e-3310-4ef3-8de1-637bf905a126/volumes" Dec 01 20:12:12 crc kubenswrapper[4960]: I1201 20:12:12.781710 4960 scope.go:117] "RemoveContainer" containerID="9c6fcea4c2b54ae1ac37450f075a146cbdbd2bc74ac325c78ed8a2707387c55c" Dec 01 20:12:12 crc kubenswrapper[4960]: I1201 20:12:12.804455 4960 scope.go:117] "RemoveContainer" containerID="e1eb5f725af2d68550a6c50a198ec70012389e94e0f620ce9871be973a497d5c" Dec 01 20:12:12 crc kubenswrapper[4960]: I1201 20:12:12.857797 4960 scope.go:117] "RemoveContainer" containerID="2b69b3a9445b35adc8c89447ecebdb94e10356d2e46bb006b364cc3ef2f92609" Dec 01 20:12:12 crc kubenswrapper[4960]: I1201 20:12:12.910961 4960 scope.go:117] "RemoveContainer" containerID="7372ee4e45a5f3cce7e2ff786d7314531acc1f0d8351fb1e1582def7a5cbc28a" Dec 01 20:12:12 crc kubenswrapper[4960]: I1201 20:12:12.971713 4960 scope.go:117] "RemoveContainer" containerID="c792bbffae1fc447430057e8a11d5f3b661d7e5ff055b189341bb1101a2d6529" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.029844 4960 scope.go:117] "RemoveContainer" containerID="d74c66b34068cfe52a3d451125cf9118482566ac828a4f6648ca6c19092f38ea" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.063668 4960 scope.go:117] "RemoveContainer" containerID="00d0dbf5e8d3f15432f74b81e5abd5cd84e4db185c5eece1c6ec9a7762218bf2" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.086458 4960 scope.go:117] "RemoveContainer" containerID="25522c134c0c582eb94b175a5405841ade1504a5a3af093d7c66c43f52b6a170" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.110584 4960 scope.go:117] "RemoveContainer" containerID="effc3da163f3ae4d02dc19621bdb0b5f2fc63641015465208e6e6ac287919008" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.132392 4960 scope.go:117] "RemoveContainer" containerID="87de426e42aba84c08c10a5023f81eb1b83297cfb0c8ae187c8fdd19c38a2b2d" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.162047 4960 scope.go:117] "RemoveContainer" containerID="f3123874e2573b16f320da0a9f0954e5a50cb3c67a9f6fff126bf8cabdf4a58f" Dec 01 20:12:13 crc kubenswrapper[4960]: I1201 20:12:13.325418 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:12:13 crc kubenswrapper[4960]: E1201 20:12:13.326174 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:12:21 crc kubenswrapper[4960]: I1201 20:12:21.046057 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qmjlr"] Dec 01 20:12:21 crc kubenswrapper[4960]: I1201 20:12:21.061101 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qmjlr"] Dec 01 20:12:21 crc kubenswrapper[4960]: I1201 20:12:21.337306 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a9d0791-d18b-4346-be7b-007e09b3add7" path="/var/lib/kubelet/pods/4a9d0791-d18b-4346-be7b-007e09b3add7/volumes" Dec 01 20:12:22 crc kubenswrapper[4960]: I1201 20:12:22.032588 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bt929"] Dec 01 20:12:22 crc kubenswrapper[4960]: I1201 20:12:22.042982 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-nd25g"] Dec 01 20:12:22 crc kubenswrapper[4960]: I1201 20:12:22.051471 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-nd25g"] Dec 01 20:12:22 crc kubenswrapper[4960]: I1201 20:12:22.064085 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bt929"] Dec 01 20:12:23 crc kubenswrapper[4960]: I1201 20:12:23.336816 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4c77e1-7c26-40f7-a01c-1902857fb5eb" path="/var/lib/kubelet/pods/2b4c77e1-7c26-40f7-a01c-1902857fb5eb/volumes" Dec 01 20:12:23 crc kubenswrapper[4960]: I1201 20:12:23.337579 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e4eb83-248f-4900-95fc-ea628cfcf067" path="/var/lib/kubelet/pods/99e4eb83-248f-4900-95fc-ea628cfcf067/volumes" Dec 01 20:12:27 crc kubenswrapper[4960]: I1201 20:12:27.324578 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:12:27 crc kubenswrapper[4960]: E1201 20:12:27.325257 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:12:35 crc kubenswrapper[4960]: I1201 20:12:35.041585 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-l6rmd"] Dec 01 20:12:35 crc kubenswrapper[4960]: I1201 20:12:35.053087 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-l6rmd"] Dec 01 20:12:35 crc kubenswrapper[4960]: I1201 20:12:35.340067 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e413a5b2-dcf3-4179-9880-b2cec1a2ef56" path="/var/lib/kubelet/pods/e413a5b2-dcf3-4179-9880-b2cec1a2ef56/volumes" Dec 01 20:12:38 crc kubenswrapper[4960]: I1201 20:12:38.324653 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:12:38 crc kubenswrapper[4960]: E1201 20:12:38.325438 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:12:53 crc kubenswrapper[4960]: I1201 20:12:53.324567 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:12:53 crc kubenswrapper[4960]: I1201 20:12:53.754442 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2"} Dec 01 20:13:13 crc kubenswrapper[4960]: I1201 20:13:13.413009 4960 scope.go:117] "RemoveContainer" containerID="733ed343a652af54cd2e8f4844a72cc95ad706f139453766e5e4e780a5728776" Dec 01 20:13:13 crc kubenswrapper[4960]: I1201 20:13:13.444926 4960 scope.go:117] "RemoveContainer" containerID="428c9e6da42206ad6467059f7724afd36582f2dc42cb5e46af4c1881182cc6d6" Dec 01 20:13:13 crc kubenswrapper[4960]: I1201 20:13:13.500627 4960 scope.go:117] "RemoveContainer" containerID="5fc63fba6074f60bae235331fdc246083cbde7e8ad97f7c9af74f179c6bf395b" Dec 01 20:13:13 crc kubenswrapper[4960]: I1201 20:13:13.553364 4960 scope.go:117] "RemoveContainer" containerID="fb5cefceb3d264c13e860d3900207851bdb4a6a3cbaf021c351a134c295c2cbc" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.073707 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8sh5f"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.086132 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0a5a-account-create-update-j28nl"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.102533 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-741e-account-create-update-pw8w9"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.112590 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ab80-account-create-update-bvzrn"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.121094 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8sh5f"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.130020 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ab80-account-create-update-bvzrn"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.139404 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-741e-account-create-update-pw8w9"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.149400 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0a5a-account-create-update-j28nl"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.159772 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-wrksp"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.168832 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-wrksp"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.176622 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jb2lh"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.183960 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jb2lh"] Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.343417 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f044d4c-9a53-4a0b-b8a8-f27eb728414b" path="/var/lib/kubelet/pods/2f044d4c-9a53-4a0b-b8a8-f27eb728414b/volumes" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.344222 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e263113-788c-4b96-bb25-5bc58b79a29a" path="/var/lib/kubelet/pods/3e263113-788c-4b96-bb25-5bc58b79a29a/volumes" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.344793 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="494b9d4e-5308-4954-92f3-39e82ff83380" path="/var/lib/kubelet/pods/494b9d4e-5308-4954-92f3-39e82ff83380/volumes" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.345348 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c257117-da7a-4840-944e-ff57c68838d6" path="/var/lib/kubelet/pods/4c257117-da7a-4840-944e-ff57c68838d6/volumes" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.346407 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98bcf1ff-8858-41e7-8c27-e0e6229b853d" path="/var/lib/kubelet/pods/98bcf1ff-8858-41e7-8c27-e0e6229b853d/volumes" Dec 01 20:13:15 crc kubenswrapper[4960]: I1201 20:13:15.346967 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac682301-3973-4d8d-9091-ab0a3e2ba851" path="/var/lib/kubelet/pods/ac682301-3973-4d8d-9091-ab0a3e2ba851/volumes" Dec 01 20:13:37 crc kubenswrapper[4960]: I1201 20:13:37.189103 4960 generic.go:334] "Generic (PLEG): container finished" podID="476f3fb9-9214-4448-9ff4-a5489883199b" containerID="eeacfd487a6cc2a4509193a76da737ce98a655e2c974ffe3a74e576191594a17" exitCode=0 Dec 01 20:13:37 crc kubenswrapper[4960]: I1201 20:13:37.189195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" event={"ID":"476f3fb9-9214-4448-9ff4-a5489883199b","Type":"ContainerDied","Data":"eeacfd487a6cc2a4509193a76da737ce98a655e2c974ffe3a74e576191594a17"} Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.667328 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.804545 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory\") pod \"476f3fb9-9214-4448-9ff4-a5489883199b\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.804987 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgdhk\" (UniqueName: \"kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk\") pod \"476f3fb9-9214-4448-9ff4-a5489883199b\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.805100 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") pod \"476f3fb9-9214-4448-9ff4-a5489883199b\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.811103 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk" (OuterVolumeSpecName: "kube-api-access-rgdhk") pod "476f3fb9-9214-4448-9ff4-a5489883199b" (UID: "476f3fb9-9214-4448-9ff4-a5489883199b"). InnerVolumeSpecName "kube-api-access-rgdhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:13:38 crc kubenswrapper[4960]: E1201 20:13:38.834289 4960 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key podName:476f3fb9-9214-4448-9ff4-a5489883199b nodeName:}" failed. No retries permitted until 2025-12-01 20:13:39.334264779 +0000 UTC m=+2054.621756448 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key" (UniqueName: "kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key") pod "476f3fb9-9214-4448-9ff4-a5489883199b" (UID: "476f3fb9-9214-4448-9ff4-a5489883199b") : error deleting /var/lib/kubelet/pods/476f3fb9-9214-4448-9ff4-a5489883199b/volume-subpaths: remove /var/lib/kubelet/pods/476f3fb9-9214-4448-9ff4-a5489883199b/volume-subpaths: no such file or directory Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.837009 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory" (OuterVolumeSpecName: "inventory") pod "476f3fb9-9214-4448-9ff4-a5489883199b" (UID: "476f3fb9-9214-4448-9ff4-a5489883199b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.907672 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgdhk\" (UniqueName: \"kubernetes.io/projected/476f3fb9-9214-4448-9ff4-a5489883199b-kube-api-access-rgdhk\") on node \"crc\" DevicePath \"\"" Dec 01 20:13:38 crc kubenswrapper[4960]: I1201 20:13:38.907711 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.216942 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" event={"ID":"476f3fb9-9214-4448-9ff4-a5489883199b","Type":"ContainerDied","Data":"6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9"} Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.216992 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bc4159cffd216e66f1990dac81400f8f835cfd56f95a78e4e5a752805e213c9" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.217044 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.312912 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt"] Dec 01 20:13:39 crc kubenswrapper[4960]: E1201 20:13:39.313582 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="476f3fb9-9214-4448-9ff4-a5489883199b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.313682 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="476f3fb9-9214-4448-9ff4-a5489883199b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.314006 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="476f3fb9-9214-4448-9ff4-a5489883199b" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.314869 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.352544 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt"] Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.416136 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") pod \"476f3fb9-9214-4448-9ff4-a5489883199b\" (UID: \"476f3fb9-9214-4448-9ff4-a5489883199b\") " Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.416747 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2rzc\" (UniqueName: \"kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.417031 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.417182 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.422271 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "476f3fb9-9214-4448-9ff4-a5489883199b" (UID: "476f3fb9-9214-4448-9ff4-a5489883199b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.518939 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.518992 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.519137 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2rzc\" (UniqueName: \"kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.519230 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/476f3fb9-9214-4448-9ff4-a5489883199b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.523345 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.523497 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.535762 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2rzc\" (UniqueName: \"kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:39 crc kubenswrapper[4960]: I1201 20:13:39.637425 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:13:40 crc kubenswrapper[4960]: I1201 20:13:40.175996 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt"] Dec 01 20:13:40 crc kubenswrapper[4960]: W1201 20:13:40.182102 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6b5eb9b_033d_4417_9289_953b2933520e.slice/crio-b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e WatchSource:0}: Error finding container b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e: Status 404 returned error can't find the container with id b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e Dec 01 20:13:40 crc kubenswrapper[4960]: I1201 20:13:40.230462 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" event={"ID":"b6b5eb9b-033d-4417-9289-953b2933520e","Type":"ContainerStarted","Data":"b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e"} Dec 01 20:13:41 crc kubenswrapper[4960]: I1201 20:13:41.241324 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" event={"ID":"b6b5eb9b-033d-4417-9289-953b2933520e","Type":"ContainerStarted","Data":"4d13659c898492a3d5e4f61ac05e41a339c1777b45ace00d922aeed6afb6b3ea"} Dec 01 20:13:41 crc kubenswrapper[4960]: I1201 20:13:41.265067 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" podStartSLOduration=1.7728495130000002 podStartE2EDuration="2.265047047s" podCreationTimestamp="2025-12-01 20:13:39 +0000 UTC" firstStartedPulling="2025-12-01 20:13:40.184685604 +0000 UTC m=+2055.472177273" lastFinishedPulling="2025-12-01 20:13:40.676883138 +0000 UTC m=+2055.964374807" observedRunningTime="2025-12-01 20:13:41.259830843 +0000 UTC m=+2056.547322512" watchObservedRunningTime="2025-12-01 20:13:41.265047047 +0000 UTC m=+2056.552538726" Dec 01 20:13:45 crc kubenswrapper[4960]: I1201 20:13:45.052245 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dknn6"] Dec 01 20:13:45 crc kubenswrapper[4960]: I1201 20:13:45.064561 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-dknn6"] Dec 01 20:13:45 crc kubenswrapper[4960]: I1201 20:13:45.337976 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0048f864-a571-46d7-9f02-ec267bd88790" path="/var/lib/kubelet/pods/0048f864-a571-46d7-9f02-ec267bd88790/volumes" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.286025 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.289325 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.308917 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.372314 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.372515 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khp46\" (UniqueName: \"kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.372781 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.474572 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khp46\" (UniqueName: \"kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.474673 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.474806 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.475649 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.475868 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.507470 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khp46\" (UniqueName: \"kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46\") pod \"redhat-marketplace-898m8\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:46 crc kubenswrapper[4960]: I1201 20:13:46.617898 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:47 crc kubenswrapper[4960]: I1201 20:13:47.131920 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:13:47 crc kubenswrapper[4960]: I1201 20:13:47.319724 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerStarted","Data":"fa37dbd01d366e0789ed61ef076f1ef3709ac03e86fb01b953abca197959e523"} Dec 01 20:13:48 crc kubenswrapper[4960]: I1201 20:13:48.333548 4960 generic.go:334] "Generic (PLEG): container finished" podID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerID="8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b" exitCode=0 Dec 01 20:13:48 crc kubenswrapper[4960]: I1201 20:13:48.333684 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerDied","Data":"8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b"} Dec 01 20:13:50 crc kubenswrapper[4960]: I1201 20:13:50.358277 4960 generic.go:334] "Generic (PLEG): container finished" podID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerID="abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53" exitCode=0 Dec 01 20:13:50 crc kubenswrapper[4960]: I1201 20:13:50.358448 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerDied","Data":"abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53"} Dec 01 20:13:51 crc kubenswrapper[4960]: I1201 20:13:51.373870 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerStarted","Data":"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836"} Dec 01 20:13:51 crc kubenswrapper[4960]: I1201 20:13:51.409725 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-898m8" podStartSLOduration=2.8682259180000003 podStartE2EDuration="5.409698284s" podCreationTimestamp="2025-12-01 20:13:46 +0000 UTC" firstStartedPulling="2025-12-01 20:13:48.336092195 +0000 UTC m=+2063.623583864" lastFinishedPulling="2025-12-01 20:13:50.877564561 +0000 UTC m=+2066.165056230" observedRunningTime="2025-12-01 20:13:51.39816807 +0000 UTC m=+2066.685659789" watchObservedRunningTime="2025-12-01 20:13:51.409698284 +0000 UTC m=+2066.697189983" Dec 01 20:13:56 crc kubenswrapper[4960]: I1201 20:13:56.618314 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:56 crc kubenswrapper[4960]: I1201 20:13:56.618688 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:56 crc kubenswrapper[4960]: I1201 20:13:56.679044 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:57 crc kubenswrapper[4960]: I1201 20:13:57.506681 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:13:57 crc kubenswrapper[4960]: I1201 20:13:57.560388 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:13:59 crc kubenswrapper[4960]: I1201 20:13:59.454727 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-898m8" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="registry-server" containerID="cri-o://9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836" gracePeriod=2 Dec 01 20:13:59 crc kubenswrapper[4960]: I1201 20:13:59.997385 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.080492 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content\") pod \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.080595 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities\") pod \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.080691 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khp46\" (UniqueName: \"kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46\") pod \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\" (UID: \"b2d68960-7fe5-4ac7-bed0-100da00c53ee\") " Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.081657 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities" (OuterVolumeSpecName: "utilities") pod "b2d68960-7fe5-4ac7-bed0-100da00c53ee" (UID: "b2d68960-7fe5-4ac7-bed0-100da00c53ee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.086940 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46" (OuterVolumeSpecName: "kube-api-access-khp46") pod "b2d68960-7fe5-4ac7-bed0-100da00c53ee" (UID: "b2d68960-7fe5-4ac7-bed0-100da00c53ee"). InnerVolumeSpecName "kube-api-access-khp46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.101868 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2d68960-7fe5-4ac7-bed0-100da00c53ee" (UID: "b2d68960-7fe5-4ac7-bed0-100da00c53ee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.183270 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.183339 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2d68960-7fe5-4ac7-bed0-100da00c53ee-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.183354 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khp46\" (UniqueName: \"kubernetes.io/projected/b2d68960-7fe5-4ac7-bed0-100da00c53ee-kube-api-access-khp46\") on node \"crc\" DevicePath \"\"" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.469807 4960 generic.go:334] "Generic (PLEG): container finished" podID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerID="9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836" exitCode=0 Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.469866 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerDied","Data":"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836"} Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.469904 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-898m8" event={"ID":"b2d68960-7fe5-4ac7-bed0-100da00c53ee","Type":"ContainerDied","Data":"fa37dbd01d366e0789ed61ef076f1ef3709ac03e86fb01b953abca197959e523"} Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.469926 4960 scope.go:117] "RemoveContainer" containerID="9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.469988 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-898m8" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.506103 4960 scope.go:117] "RemoveContainer" containerID="abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.512738 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.522737 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-898m8"] Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.535846 4960 scope.go:117] "RemoveContainer" containerID="8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.594372 4960 scope.go:117] "RemoveContainer" containerID="9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836" Dec 01 20:14:00 crc kubenswrapper[4960]: E1201 20:14:00.594943 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836\": container with ID starting with 9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836 not found: ID does not exist" containerID="9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.594984 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836"} err="failed to get container status \"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836\": rpc error: code = NotFound desc = could not find container \"9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836\": container with ID starting with 9968522f21a12264decf92604de3e3059b3e643f622bacc80cce47777ee71836 not found: ID does not exist" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.595014 4960 scope.go:117] "RemoveContainer" containerID="abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53" Dec 01 20:14:00 crc kubenswrapper[4960]: E1201 20:14:00.595445 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53\": container with ID starting with abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53 not found: ID does not exist" containerID="abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.595512 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53"} err="failed to get container status \"abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53\": rpc error: code = NotFound desc = could not find container \"abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53\": container with ID starting with abaebae804b3cd2948034e222a03f25a097b6cacf36a9ac7a2e97a80a2f91b53 not found: ID does not exist" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.595573 4960 scope.go:117] "RemoveContainer" containerID="8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b" Dec 01 20:14:00 crc kubenswrapper[4960]: E1201 20:14:00.595977 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b\": container with ID starting with 8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b not found: ID does not exist" containerID="8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b" Dec 01 20:14:00 crc kubenswrapper[4960]: I1201 20:14:00.596018 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b"} err="failed to get container status \"8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b\": rpc error: code = NotFound desc = could not find container \"8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b\": container with ID starting with 8abeabce81f526fed5af40274d90daa7e09a768bfd2cb66cb093cb4ac99d109b not found: ID does not exist" Dec 01 20:14:01 crc kubenswrapper[4960]: I1201 20:14:01.342892 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" path="/var/lib/kubelet/pods/b2d68960-7fe5-4ac7-bed0-100da00c53ee/volumes" Dec 01 20:14:08 crc kubenswrapper[4960]: I1201 20:14:08.062507 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-8mhss"] Dec 01 20:14:08 crc kubenswrapper[4960]: I1201 20:14:08.074175 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-8mhss"] Dec 01 20:14:09 crc kubenswrapper[4960]: I1201 20:14:09.042814 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvgdt"] Dec 01 20:14:09 crc kubenswrapper[4960]: I1201 20:14:09.056273 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tvgdt"] Dec 01 20:14:09 crc kubenswrapper[4960]: I1201 20:14:09.338539 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dae60dd-ebab-42c1-9dc6-722768140fc4" path="/var/lib/kubelet/pods/5dae60dd-ebab-42c1-9dc6-722768140fc4/volumes" Dec 01 20:14:09 crc kubenswrapper[4960]: I1201 20:14:09.339233 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82af32f2-8438-4070-814e-bfad17fe42b9" path="/var/lib/kubelet/pods/82af32f2-8438-4070-814e-bfad17fe42b9/volumes" Dec 01 20:14:13 crc kubenswrapper[4960]: I1201 20:14:13.697312 4960 scope.go:117] "RemoveContainer" containerID="80b33f81a5841f3ec676a4a386b16a7c5b56e34fa76b76bcda3766b726a439d1" Dec 01 20:14:13 crc kubenswrapper[4960]: I1201 20:14:13.746869 4960 scope.go:117] "RemoveContainer" containerID="42d4361844a02929c43bfd8fbca28a5c7baba6d859ae982a6ce92b0d44070742" Dec 01 20:14:13 crc kubenswrapper[4960]: I1201 20:14:13.802028 4960 scope.go:117] "RemoveContainer" containerID="509edc4faa04faee93bd0e25d7b0404ded6224148eba909a1a0690edf60e7a26" Dec 01 20:14:13 crc kubenswrapper[4960]: I1201 20:14:13.888426 4960 scope.go:117] "RemoveContainer" containerID="d6d87e3675c17909fba324aaad98765ff94ec8c0a15823449f65c7edd7f4e5ba" Dec 01 20:14:13 crc kubenswrapper[4960]: I1201 20:14:13.926286 4960 scope.go:117] "RemoveContainer" containerID="f3c7043a62c3cb27566e56f287428770bb3c204811b417c181879cc68a3e84ec" Dec 01 20:14:14 crc kubenswrapper[4960]: I1201 20:14:14.008652 4960 scope.go:117] "RemoveContainer" containerID="8be31e18d2bc3a186cafd70c832374703a971d9d22da5cde6ca11ce540f4d38e" Dec 01 20:14:14 crc kubenswrapper[4960]: I1201 20:14:14.035416 4960 scope.go:117] "RemoveContainer" containerID="f0b0f2e0596daff6ec7c5ad7f83a5b8d7c159d26338123d3e8ab1361d21f9c4d" Dec 01 20:14:14 crc kubenswrapper[4960]: I1201 20:14:14.060109 4960 scope.go:117] "RemoveContainer" containerID="76c1e247963966f85a2bbde937758f7c67f0ed6a5dbc5ab1cd889fff29d08b07" Dec 01 20:14:14 crc kubenswrapper[4960]: I1201 20:14:14.084647 4960 scope.go:117] "RemoveContainer" containerID="d114c4107a031ac1b3d9ac0420107c3bd595c0d075c5c6ba58438a416c43acc6" Dec 01 20:14:54 crc kubenswrapper[4960]: I1201 20:14:54.055260 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-tvqfm"] Dec 01 20:14:54 crc kubenswrapper[4960]: I1201 20:14:54.066194 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-tvqfm"] Dec 01 20:14:55 crc kubenswrapper[4960]: I1201 20:14:55.336851 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9445a1ff-2038-490e-9bf2-31d038d79ceb" path="/var/lib/kubelet/pods/9445a1ff-2038-490e-9bf2-31d038d79ceb/volumes" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.160360 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf"] Dec 01 20:15:00 crc kubenswrapper[4960]: E1201 20:15:00.161482 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="registry-server" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.161502 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="registry-server" Dec 01 20:15:00 crc kubenswrapper[4960]: E1201 20:15:00.161520 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="extract-content" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.161528 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="extract-content" Dec 01 20:15:00 crc kubenswrapper[4960]: E1201 20:15:00.161539 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="extract-utilities" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.161549 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="extract-utilities" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.161811 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d68960-7fe5-4ac7-bed0-100da00c53ee" containerName="registry-server" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.162706 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.164399 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.164430 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.170766 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf"] Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.297480 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.297597 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj8gt\" (UniqueName: \"kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.297627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.399844 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.399974 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj8gt\" (UniqueName: \"kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.400013 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.401184 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.407185 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.427568 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj8gt\" (UniqueName: \"kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt\") pod \"collect-profiles-29410335-4lkbf\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.490954 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:00 crc kubenswrapper[4960]: I1201 20:15:00.940216 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf"] Dec 01 20:15:01 crc kubenswrapper[4960]: I1201 20:15:01.121303 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" event={"ID":"9f4a047e-d49a-4f90-867e-5c04f2c572fc","Type":"ContainerStarted","Data":"8d67f6b4fe7757bc7a98ee71fbc0b6d0b44e3b0b4741ccfd5d4219dce1ccb323"} Dec 01 20:15:02 crc kubenswrapper[4960]: I1201 20:15:02.131584 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6b5eb9b-033d-4417-9289-953b2933520e" containerID="4d13659c898492a3d5e4f61ac05e41a339c1777b45ace00d922aeed6afb6b3ea" exitCode=0 Dec 01 20:15:02 crc kubenswrapper[4960]: I1201 20:15:02.131799 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" event={"ID":"b6b5eb9b-033d-4417-9289-953b2933520e","Type":"ContainerDied","Data":"4d13659c898492a3d5e4f61ac05e41a339c1777b45ace00d922aeed6afb6b3ea"} Dec 01 20:15:02 crc kubenswrapper[4960]: I1201 20:15:02.133793 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" event={"ID":"9f4a047e-d49a-4f90-867e-5c04f2c572fc","Type":"ContainerStarted","Data":"92d37c935b5f06e9c2b65a080f9612efaa5d5abd3694db85c36917831ece1f20"} Dec 01 20:15:02 crc kubenswrapper[4960]: I1201 20:15:02.206057 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" podStartSLOduration=2.206034263 podStartE2EDuration="2.206034263s" podCreationTimestamp="2025-12-01 20:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:15:02.189232028 +0000 UTC m=+2137.476723717" watchObservedRunningTime="2025-12-01 20:15:02.206034263 +0000 UTC m=+2137.493525932" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.145233 4960 generic.go:334] "Generic (PLEG): container finished" podID="9f4a047e-d49a-4f90-867e-5c04f2c572fc" containerID="92d37c935b5f06e9c2b65a080f9612efaa5d5abd3694db85c36917831ece1f20" exitCode=0 Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.145337 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" event={"ID":"9f4a047e-d49a-4f90-867e-5c04f2c572fc","Type":"ContainerDied","Data":"92d37c935b5f06e9c2b65a080f9612efaa5d5abd3694db85c36917831ece1f20"} Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.634256 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.769380 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2rzc\" (UniqueName: \"kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc\") pod \"b6b5eb9b-033d-4417-9289-953b2933520e\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.769464 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key\") pod \"b6b5eb9b-033d-4417-9289-953b2933520e\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.769585 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory\") pod \"b6b5eb9b-033d-4417-9289-953b2933520e\" (UID: \"b6b5eb9b-033d-4417-9289-953b2933520e\") " Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.775637 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc" (OuterVolumeSpecName: "kube-api-access-n2rzc") pod "b6b5eb9b-033d-4417-9289-953b2933520e" (UID: "b6b5eb9b-033d-4417-9289-953b2933520e"). InnerVolumeSpecName "kube-api-access-n2rzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.799505 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6b5eb9b-033d-4417-9289-953b2933520e" (UID: "b6b5eb9b-033d-4417-9289-953b2933520e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.804220 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory" (OuterVolumeSpecName: "inventory") pod "b6b5eb9b-033d-4417-9289-953b2933520e" (UID: "b6b5eb9b-033d-4417-9289-953b2933520e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.872365 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.872412 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2rzc\" (UniqueName: \"kubernetes.io/projected/b6b5eb9b-033d-4417-9289-953b2933520e-kube-api-access-n2rzc\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:03 crc kubenswrapper[4960]: I1201 20:15:03.872430 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6b5eb9b-033d-4417-9289-953b2933520e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.157619 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.157672 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt" event={"ID":"b6b5eb9b-033d-4417-9289-953b2933520e","Type":"ContainerDied","Data":"b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e"} Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.157701 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b660121ef83da5d6ca65a08ed4e34ce54714c2538da8adaffffbe24ec5811b8e" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.241356 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8"] Dec 01 20:15:04 crc kubenswrapper[4960]: E1201 20:15:04.241821 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b5eb9b-033d-4417-9289-953b2933520e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.241841 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b5eb9b-033d-4417-9289-953b2933520e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.242054 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b5eb9b-033d-4417-9289-953b2933520e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.242923 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.246355 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.246935 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.247329 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.247598 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.265253 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8"] Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.385629 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdxzq\" (UniqueName: \"kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.385787 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.385923 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.487780 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.487908 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdxzq\" (UniqueName: \"kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.487963 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.492890 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.496446 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.504759 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdxzq\" (UniqueName: \"kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.584307 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.710995 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.898654 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume\") pod \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.898777 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume\") pod \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.898870 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj8gt\" (UniqueName: \"kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt\") pod \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\" (UID: \"9f4a047e-d49a-4f90-867e-5c04f2c572fc\") " Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.899567 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume" (OuterVolumeSpecName: "config-volume") pod "9f4a047e-d49a-4f90-867e-5c04f2c572fc" (UID: "9f4a047e-d49a-4f90-867e-5c04f2c572fc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.903551 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt" (OuterVolumeSpecName: "kube-api-access-zj8gt") pod "9f4a047e-d49a-4f90-867e-5c04f2c572fc" (UID: "9f4a047e-d49a-4f90-867e-5c04f2c572fc"). InnerVolumeSpecName "kube-api-access-zj8gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:04 crc kubenswrapper[4960]: I1201 20:15:04.904325 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9f4a047e-d49a-4f90-867e-5c04f2c572fc" (UID: "9f4a047e-d49a-4f90-867e-5c04f2c572fc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.001023 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f4a047e-d49a-4f90-867e-5c04f2c572fc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.001325 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj8gt\" (UniqueName: \"kubernetes.io/projected/9f4a047e-d49a-4f90-867e-5c04f2c572fc-kube-api-access-zj8gt\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.001404 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f4a047e-d49a-4f90-867e-5c04f2c572fc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.119166 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8"] Dec 01 20:15:05 crc kubenswrapper[4960]: W1201 20:15:05.123889 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8dcd369_d6cb_4472_8530_9c4d58ac2a0c.slice/crio-2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6 WatchSource:0}: Error finding container 2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6: Status 404 returned error can't find the container with id 2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6 Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.167848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" event={"ID":"9f4a047e-d49a-4f90-867e-5c04f2c572fc","Type":"ContainerDied","Data":"8d67f6b4fe7757bc7a98ee71fbc0b6d0b44e3b0b4741ccfd5d4219dce1ccb323"} Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.167888 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d67f6b4fe7757bc7a98ee71fbc0b6d0b44e3b0b4741ccfd5d4219dce1ccb323" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.167913 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf" Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.172409 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" event={"ID":"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c","Type":"ContainerStarted","Data":"2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6"} Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.785947 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb"] Dec 01 20:15:05 crc kubenswrapper[4960]: I1201 20:15:05.798758 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410290-c9bfb"] Dec 01 20:15:06 crc kubenswrapper[4960]: I1201 20:15:06.184772 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" event={"ID":"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c","Type":"ContainerStarted","Data":"68a64aa4c8a5d05c927d6936529374fc793c930da383859d1f3126480a6cb7ee"} Dec 01 20:15:06 crc kubenswrapper[4960]: I1201 20:15:06.206658 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" podStartSLOduration=1.650831038 podStartE2EDuration="2.206637173s" podCreationTimestamp="2025-12-01 20:15:04 +0000 UTC" firstStartedPulling="2025-12-01 20:15:05.126198058 +0000 UTC m=+2140.413689747" lastFinishedPulling="2025-12-01 20:15:05.682004203 +0000 UTC m=+2140.969495882" observedRunningTime="2025-12-01 20:15:06.20013548 +0000 UTC m=+2141.487627149" watchObservedRunningTime="2025-12-01 20:15:06.206637173 +0000 UTC m=+2141.494128842" Dec 01 20:15:07 crc kubenswrapper[4960]: I1201 20:15:07.351085 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c430c72-691c-4a6c-b4c5-9edb65fc8a31" path="/var/lib/kubelet/pods/3c430c72-691c-4a6c-b4c5-9edb65fc8a31/volumes" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.338561 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:08 crc kubenswrapper[4960]: E1201 20:15:08.339152 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4a047e-d49a-4f90-867e-5c04f2c572fc" containerName="collect-profiles" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.339173 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4a047e-d49a-4f90-867e-5c04f2c572fc" containerName="collect-profiles" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.339605 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4a047e-d49a-4f90-867e-5c04f2c572fc" containerName="collect-profiles" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.341580 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.350631 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.476733 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.476819 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.476982 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4mk4\" (UniqueName: \"kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.579456 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4mk4\" (UniqueName: \"kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.579620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.579679 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.580179 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.580210 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.601853 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4mk4\" (UniqueName: \"kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4\") pod \"certified-operators-qtws8\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:08 crc kubenswrapper[4960]: I1201 20:15:08.663419 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:09 crc kubenswrapper[4960]: I1201 20:15:09.309257 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.222581 4960 generic.go:334] "Generic (PLEG): container finished" podID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerID="8fb0d9b87eafcdd7d91ef25497468bef1d9d02e9bfc378ed8b81279a2aee8b54" exitCode=0 Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.222664 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerDied","Data":"8fb0d9b87eafcdd7d91ef25497468bef1d9d02e9bfc378ed8b81279a2aee8b54"} Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.222859 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerStarted","Data":"9de28cc7974ea9dea97b48f67b479d35105e389b146b912525df4ff7228545d0"} Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.750032 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.752561 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.775195 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.928712 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.929010 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:10 crc kubenswrapper[4960]: I1201 20:15:10.929072 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvfnf\" (UniqueName: \"kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.030776 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.030884 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvfnf\" (UniqueName: \"kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.030975 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.031618 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.031715 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.058513 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvfnf\" (UniqueName: \"kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf\") pod \"redhat-operators-2psm7\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.090982 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.239974 4960 generic.go:334] "Generic (PLEG): container finished" podID="d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" containerID="68a64aa4c8a5d05c927d6936529374fc793c930da383859d1f3126480a6cb7ee" exitCode=0 Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.240167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" event={"ID":"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c","Type":"ContainerDied","Data":"68a64aa4c8a5d05c927d6936529374fc793c930da383859d1f3126480a6cb7ee"} Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.247140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerStarted","Data":"29f31d3c46bbaea0f4acf5785f2445743536e6e0916e9a3bafc22a5789c07dc4"} Dec 01 20:15:11 crc kubenswrapper[4960]: I1201 20:15:11.592032 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:12 crc kubenswrapper[4960]: I1201 20:15:12.261292 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerStarted","Data":"e7971ec85650eecca33af18e9663429ec11d2811893870039f553be274ce37bd"} Dec 01 20:15:12 crc kubenswrapper[4960]: I1201 20:15:12.812341 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:12 crc kubenswrapper[4960]: I1201 20:15:12.978944 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdxzq\" (UniqueName: \"kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq\") pod \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " Dec 01 20:15:12 crc kubenswrapper[4960]: I1201 20:15:12.979129 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key\") pod \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " Dec 01 20:15:12 crc kubenswrapper[4960]: I1201 20:15:12.979213 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory\") pod \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\" (UID: \"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c\") " Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.027380 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq" (OuterVolumeSpecName: "kube-api-access-hdxzq") pod "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" (UID: "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c"). InnerVolumeSpecName "kube-api-access-hdxzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.048468 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" (UID: "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.081688 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdxzq\" (UniqueName: \"kubernetes.io/projected/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-kube-api-access-hdxzq\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.081736 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.090670 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory" (OuterVolumeSpecName: "inventory") pod "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" (UID: "d8dcd369-d6cb-4472-8530-9c4d58ac2a0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.184364 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8dcd369-d6cb-4472-8530-9c4d58ac2a0c-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.269977 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" event={"ID":"d8dcd369-d6cb-4472-8530-9c4d58ac2a0c","Type":"ContainerDied","Data":"2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6"} Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.270015 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ba5a57eaf5b5d72186d6321ab13802a7389d4f1f4580481f8f64d9041fe52c6" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.270027 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.538917 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x"] Dec 01 20:15:13 crc kubenswrapper[4960]: E1201 20:15:13.539821 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.539846 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.540152 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8dcd369-d6cb-4472-8530-9c4d58ac2a0c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.541167 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.544796 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.545033 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.548674 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.553638 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.564224 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x"] Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.694380 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64t5m\" (UniqueName: \"kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.694431 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.694470 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.796992 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.797059 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.797304 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64t5m\" (UniqueName: \"kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.810761 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.814498 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.821860 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64t5m\" (UniqueName: \"kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-9bb2x\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.863990 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.870957 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:15:13 crc kubenswrapper[4960]: I1201 20:15:13.871022 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:15:14 crc kubenswrapper[4960]: I1201 20:15:14.284753 4960 generic.go:334] "Generic (PLEG): container finished" podID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerID="29f31d3c46bbaea0f4acf5785f2445743536e6e0916e9a3bafc22a5789c07dc4" exitCode=0 Dec 01 20:15:14 crc kubenswrapper[4960]: I1201 20:15:14.284834 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerDied","Data":"29f31d3c46bbaea0f4acf5785f2445743536e6e0916e9a3bafc22a5789c07dc4"} Dec 01 20:15:14 crc kubenswrapper[4960]: I1201 20:15:14.298859 4960 scope.go:117] "RemoveContainer" containerID="fb65c6f74e217e163c05263be53dfb9dd14f26f56b08f9429f1b94ee5c2cc8a8" Dec 01 20:15:14 crc kubenswrapper[4960]: I1201 20:15:14.420409 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x"] Dec 01 20:15:14 crc kubenswrapper[4960]: W1201 20:15:14.428767 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aa6823d_0480_490a_977b_801110a10181.slice/crio-05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d WatchSource:0}: Error finding container 05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d: Status 404 returned error can't find the container with id 05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d Dec 01 20:15:15 crc kubenswrapper[4960]: I1201 20:15:15.071459 4960 scope.go:117] "RemoveContainer" containerID="65d842ea69cd7300ac4915f111b07672025c584563b99e88dafcac489c55f999" Dec 01 20:15:15 crc kubenswrapper[4960]: I1201 20:15:15.300751 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" event={"ID":"4aa6823d-0480-490a-977b-801110a10181","Type":"ContainerStarted","Data":"05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d"} Dec 01 20:15:15 crc kubenswrapper[4960]: I1201 20:15:15.305097 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerStarted","Data":"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2"} Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.316621 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" event={"ID":"4aa6823d-0480-490a-977b-801110a10181","Type":"ContainerStarted","Data":"7735832ac57eda6b9491fbea03c25460054b9be3104cb7201982c4e5c0dcf540"} Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.321581 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerStarted","Data":"81c1acf876d27c11059b9a2d0fabd1479bde0e97b7c3561b42b6d598536e8dbe"} Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.323606 4960 generic.go:334] "Generic (PLEG): container finished" podID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerID="fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2" exitCode=0 Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.323651 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerDied","Data":"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2"} Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.343203 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" podStartSLOduration=1.924617161 podStartE2EDuration="3.343184236s" podCreationTimestamp="2025-12-01 20:15:13 +0000 UTC" firstStartedPulling="2025-12-01 20:15:14.43128966 +0000 UTC m=+2149.718781329" lastFinishedPulling="2025-12-01 20:15:15.849856735 +0000 UTC m=+2151.137348404" observedRunningTime="2025-12-01 20:15:16.334456403 +0000 UTC m=+2151.621948082" watchObservedRunningTime="2025-12-01 20:15:16.343184236 +0000 UTC m=+2151.630675905" Dec 01 20:15:16 crc kubenswrapper[4960]: I1201 20:15:16.383358 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qtws8" podStartSLOduration=2.919420798 podStartE2EDuration="8.383333112s" podCreationTimestamp="2025-12-01 20:15:08 +0000 UTC" firstStartedPulling="2025-12-01 20:15:10.224985179 +0000 UTC m=+2145.512476848" lastFinishedPulling="2025-12-01 20:15:15.688897493 +0000 UTC m=+2150.976389162" observedRunningTime="2025-12-01 20:15:16.376322412 +0000 UTC m=+2151.663814091" watchObservedRunningTime="2025-12-01 20:15:16.383333112 +0000 UTC m=+2151.670824791" Dec 01 20:15:18 crc kubenswrapper[4960]: I1201 20:15:18.349849 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerStarted","Data":"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead"} Dec 01 20:15:18 crc kubenswrapper[4960]: I1201 20:15:18.664665 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:18 crc kubenswrapper[4960]: I1201 20:15:18.664749 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:18 crc kubenswrapper[4960]: I1201 20:15:18.723976 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:21 crc kubenswrapper[4960]: I1201 20:15:21.380424 4960 generic.go:334] "Generic (PLEG): container finished" podID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerID="e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead" exitCode=0 Dec 01 20:15:21 crc kubenswrapper[4960]: I1201 20:15:21.381104 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerDied","Data":"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead"} Dec 01 20:15:22 crc kubenswrapper[4960]: I1201 20:15:22.393876 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerStarted","Data":"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b"} Dec 01 20:15:22 crc kubenswrapper[4960]: I1201 20:15:22.415136 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2psm7" podStartSLOduration=6.852053205 podStartE2EDuration="12.415095557s" podCreationTimestamp="2025-12-01 20:15:10 +0000 UTC" firstStartedPulling="2025-12-01 20:15:16.325591917 +0000 UTC m=+2151.613083586" lastFinishedPulling="2025-12-01 20:15:21.888634259 +0000 UTC m=+2157.176125938" observedRunningTime="2025-12-01 20:15:22.407919313 +0000 UTC m=+2157.695410982" watchObservedRunningTime="2025-12-01 20:15:22.415095557 +0000 UTC m=+2157.702587216" Dec 01 20:15:25 crc kubenswrapper[4960]: I1201 20:15:25.948556 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:25 crc kubenswrapper[4960]: I1201 20:15:25.951405 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:25 crc kubenswrapper[4960]: I1201 20:15:25.965041 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.070719 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.070835 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcv8x\" (UniqueName: \"kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.070918 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.172283 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.172449 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.172503 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcv8x\" (UniqueName: \"kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.172954 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.172961 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.199503 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcv8x\" (UniqueName: \"kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x\") pod \"community-operators-77s9l\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:26 crc kubenswrapper[4960]: I1201 20:15:26.269702 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:27 crc kubenswrapper[4960]: I1201 20:15:27.165525 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:27 crc kubenswrapper[4960]: W1201 20:15:27.167669 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0043e5b_1732_44f2_8afe_4c5a2e52c130.slice/crio-df7741781cfa3d6722d52dc6ba9134c007ee124e11af01f747d3759235a18fcc WatchSource:0}: Error finding container df7741781cfa3d6722d52dc6ba9134c007ee124e11af01f747d3759235a18fcc: Status 404 returned error can't find the container with id df7741781cfa3d6722d52dc6ba9134c007ee124e11af01f747d3759235a18fcc Dec 01 20:15:27 crc kubenswrapper[4960]: I1201 20:15:27.443224 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerID="b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9" exitCode=0 Dec 01 20:15:27 crc kubenswrapper[4960]: I1201 20:15:27.443285 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerDied","Data":"b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9"} Dec 01 20:15:27 crc kubenswrapper[4960]: I1201 20:15:27.443500 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerStarted","Data":"df7741781cfa3d6722d52dc6ba9134c007ee124e11af01f747d3759235a18fcc"} Dec 01 20:15:28 crc kubenswrapper[4960]: I1201 20:15:28.749025 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:29 crc kubenswrapper[4960]: I1201 20:15:29.465751 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerStarted","Data":"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7"} Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.342852 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.343146 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qtws8" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="registry-server" containerID="cri-o://81c1acf876d27c11059b9a2d0fabd1479bde0e97b7c3561b42b6d598536e8dbe" gracePeriod=2 Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.483585 4960 generic.go:334] "Generic (PLEG): container finished" podID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerID="81c1acf876d27c11059b9a2d0fabd1479bde0e97b7c3561b42b6d598536e8dbe" exitCode=0 Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.483658 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerDied","Data":"81c1acf876d27c11059b9a2d0fabd1479bde0e97b7c3561b42b6d598536e8dbe"} Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.486770 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerID="3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7" exitCode=0 Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.486827 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerDied","Data":"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7"} Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.927078 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.999002 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content\") pod \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.999557 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities\") pod \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " Dec 01 20:15:30 crc kubenswrapper[4960]: I1201 20:15:30.999729 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4mk4\" (UniqueName: \"kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4\") pod \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\" (UID: \"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6\") " Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.001461 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities" (OuterVolumeSpecName: "utilities") pod "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" (UID: "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.005708 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.027956 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4" (OuterVolumeSpecName: "kube-api-access-b4mk4") pod "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" (UID: "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6"). InnerVolumeSpecName "kube-api-access-b4mk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.080942 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" (UID: "79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.091685 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.092090 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.107699 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.107732 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4mk4\" (UniqueName: \"kubernetes.io/projected/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6-kube-api-access-b4mk4\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.141791 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.497871 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qtws8" event={"ID":"79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6","Type":"ContainerDied","Data":"9de28cc7974ea9dea97b48f67b479d35105e389b146b912525df4ff7228545d0"} Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.498304 4960 scope.go:117] "RemoveContainer" containerID="81c1acf876d27c11059b9a2d0fabd1479bde0e97b7c3561b42b6d598536e8dbe" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.497899 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qtws8" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.525833 4960 scope.go:117] "RemoveContainer" containerID="29f31d3c46bbaea0f4acf5785f2445743536e6e0916e9a3bafc22a5789c07dc4" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.527041 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.545793 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qtws8"] Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.556221 4960 scope.go:117] "RemoveContainer" containerID="8fb0d9b87eafcdd7d91ef25497468bef1d9d02e9bfc378ed8b81279a2aee8b54" Dec 01 20:15:31 crc kubenswrapper[4960]: I1201 20:15:31.558791 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:32 crc kubenswrapper[4960]: I1201 20:15:32.510153 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerStarted","Data":"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354"} Dec 01 20:15:32 crc kubenswrapper[4960]: I1201 20:15:32.538460 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-77s9l" podStartSLOduration=3.666490988 podStartE2EDuration="7.538439176s" podCreationTimestamp="2025-12-01 20:15:25 +0000 UTC" firstStartedPulling="2025-12-01 20:15:27.446556063 +0000 UTC m=+2162.734047742" lastFinishedPulling="2025-12-01 20:15:31.318504221 +0000 UTC m=+2166.605995930" observedRunningTime="2025-12-01 20:15:32.530291782 +0000 UTC m=+2167.817783471" watchObservedRunningTime="2025-12-01 20:15:32.538439176 +0000 UTC m=+2167.825930845" Dec 01 20:15:33 crc kubenswrapper[4960]: I1201 20:15:33.341366 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" path="/var/lib/kubelet/pods/79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6/volumes" Dec 01 20:15:33 crc kubenswrapper[4960]: I1201 20:15:33.737570 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:34 crc kubenswrapper[4960]: I1201 20:15:34.531744 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2psm7" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="registry-server" containerID="cri-o://96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b" gracePeriod=2 Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.070062 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.200628 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities\") pod \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.200855 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvfnf\" (UniqueName: \"kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf\") pod \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.201027 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content\") pod \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\" (UID: \"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f\") " Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.204644 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities" (OuterVolumeSpecName: "utilities") pod "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" (UID: "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.213427 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf" (OuterVolumeSpecName: "kube-api-access-jvfnf") pod "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" (UID: "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f"). InnerVolumeSpecName "kube-api-access-jvfnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.306592 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.306658 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvfnf\" (UniqueName: \"kubernetes.io/projected/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-kube-api-access-jvfnf\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.347144 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" (UID: "bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.408431 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.543316 4960 generic.go:334] "Generic (PLEG): container finished" podID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerID="96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b" exitCode=0 Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.543368 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerDied","Data":"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b"} Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.543426 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2psm7" event={"ID":"bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f","Type":"ContainerDied","Data":"e7971ec85650eecca33af18e9663429ec11d2811893870039f553be274ce37bd"} Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.543432 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2psm7" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.543455 4960 scope.go:117] "RemoveContainer" containerID="96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.570640 4960 scope.go:117] "RemoveContainer" containerID="e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.591666 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.602182 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2psm7"] Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.602372 4960 scope.go:117] "RemoveContainer" containerID="fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.647147 4960 scope.go:117] "RemoveContainer" containerID="96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b" Dec 01 20:15:35 crc kubenswrapper[4960]: E1201 20:15:35.647603 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b\": container with ID starting with 96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b not found: ID does not exist" containerID="96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.647656 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b"} err="failed to get container status \"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b\": rpc error: code = NotFound desc = could not find container \"96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b\": container with ID starting with 96b87a21cbb7e3b739a916c4847c45b6c005331489c90c2016ec944d39f26c6b not found: ID does not exist" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.647692 4960 scope.go:117] "RemoveContainer" containerID="e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead" Dec 01 20:15:35 crc kubenswrapper[4960]: E1201 20:15:35.648465 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead\": container with ID starting with e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead not found: ID does not exist" containerID="e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.648490 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead"} err="failed to get container status \"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead\": rpc error: code = NotFound desc = could not find container \"e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead\": container with ID starting with e51ae61470a52abd7fd3a710a592c6ccd47e8effa5564083778db11629487ead not found: ID does not exist" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.648511 4960 scope.go:117] "RemoveContainer" containerID="fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2" Dec 01 20:15:35 crc kubenswrapper[4960]: E1201 20:15:35.649301 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2\": container with ID starting with fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2 not found: ID does not exist" containerID="fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2" Dec 01 20:15:35 crc kubenswrapper[4960]: I1201 20:15:35.649346 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2"} err="failed to get container status \"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2\": rpc error: code = NotFound desc = could not find container \"fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2\": container with ID starting with fb77f1ce4a2b64ee73daf943775e2024036d068e0595f15b62fba71da61d70c2 not found: ID does not exist" Dec 01 20:15:36 crc kubenswrapper[4960]: I1201 20:15:36.271406 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:36 crc kubenswrapper[4960]: I1201 20:15:36.272047 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:36 crc kubenswrapper[4960]: I1201 20:15:36.338626 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:36 crc kubenswrapper[4960]: I1201 20:15:36.603760 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:37 crc kubenswrapper[4960]: I1201 20:15:37.337677 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" path="/var/lib/kubelet/pods/bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f/volumes" Dec 01 20:15:38 crc kubenswrapper[4960]: I1201 20:15:38.062320 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-db-sync-dl272"] Dec 01 20:15:38 crc kubenswrapper[4960]: I1201 20:15:38.080364 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-db-sync-dl272"] Dec 01 20:15:38 crc kubenswrapper[4960]: I1201 20:15:38.135027 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:38 crc kubenswrapper[4960]: I1201 20:15:38.570553 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-77s9l" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="registry-server" containerID="cri-o://21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354" gracePeriod=2 Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.283417 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.396141 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities\") pod \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.396191 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcv8x\" (UniqueName: \"kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x\") pod \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.396405 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content\") pod \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\" (UID: \"e0043e5b-1732-44f2-8afe-4c5a2e52c130\") " Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.400564 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities" (OuterVolumeSpecName: "utilities") pod "e0043e5b-1732-44f2-8afe-4c5a2e52c130" (UID: "e0043e5b-1732-44f2-8afe-4c5a2e52c130"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.423892 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x" (OuterVolumeSpecName: "kube-api-access-pcv8x") pod "e0043e5b-1732-44f2-8afe-4c5a2e52c130" (UID: "e0043e5b-1732-44f2-8afe-4c5a2e52c130"). InnerVolumeSpecName "kube-api-access-pcv8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.425257 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f838587c-b056-4d32-bf13-e6a7a0c8da50" path="/var/lib/kubelet/pods/f838587c-b056-4d32-bf13-e6a7a0c8da50/volumes" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.494395 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0043e5b-1732-44f2-8afe-4c5a2e52c130" (UID: "e0043e5b-1732-44f2-8afe-4c5a2e52c130"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.499947 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.500997 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0043e5b-1732-44f2-8afe-4c5a2e52c130-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.501011 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcv8x\" (UniqueName: \"kubernetes.io/projected/e0043e5b-1732-44f2-8afe-4c5a2e52c130-kube-api-access-pcv8x\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.582877 4960 generic.go:334] "Generic (PLEG): container finished" podID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerID="21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354" exitCode=0 Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.582930 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerDied","Data":"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354"} Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.582956 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-77s9l" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.582977 4960 scope.go:117] "RemoveContainer" containerID="21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.582964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-77s9l" event={"ID":"e0043e5b-1732-44f2-8afe-4c5a2e52c130","Type":"ContainerDied","Data":"df7741781cfa3d6722d52dc6ba9134c007ee124e11af01f747d3759235a18fcc"} Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.614656 4960 scope.go:117] "RemoveContainer" containerID="3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.622700 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.639819 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-77s9l"] Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.651054 4960 scope.go:117] "RemoveContainer" containerID="b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.696204 4960 scope.go:117] "RemoveContainer" containerID="21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354" Dec 01 20:15:39 crc kubenswrapper[4960]: E1201 20:15:39.696911 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354\": container with ID starting with 21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354 not found: ID does not exist" containerID="21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.697006 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354"} err="failed to get container status \"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354\": rpc error: code = NotFound desc = could not find container \"21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354\": container with ID starting with 21146aeb80b02c60d946a454ed561fbfdadcb69d9bf2c4a329c21562ee106354 not found: ID does not exist" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.697086 4960 scope.go:117] "RemoveContainer" containerID="3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7" Dec 01 20:15:39 crc kubenswrapper[4960]: E1201 20:15:39.697471 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7\": container with ID starting with 3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7 not found: ID does not exist" containerID="3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.697494 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7"} err="failed to get container status \"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7\": rpc error: code = NotFound desc = could not find container \"3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7\": container with ID starting with 3c1152db12d0c97965e86f64a06d9f3ce36e506038d83b80ad8f77a49576a6d7 not found: ID does not exist" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.697509 4960 scope.go:117] "RemoveContainer" containerID="b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9" Dec 01 20:15:39 crc kubenswrapper[4960]: E1201 20:15:39.697756 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9\": container with ID starting with b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9 not found: ID does not exist" containerID="b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9" Dec 01 20:15:39 crc kubenswrapper[4960]: I1201 20:15:39.697852 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9"} err="failed to get container status \"b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9\": rpc error: code = NotFound desc = could not find container \"b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9\": container with ID starting with b886cbf847d2b57bfcbb62a7d7a882ae71f882664338797334734572c74372d9 not found: ID does not exist" Dec 01 20:15:41 crc kubenswrapper[4960]: I1201 20:15:41.338772 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" path="/var/lib/kubelet/pods/e0043e5b-1732-44f2-8afe-4c5a2e52c130/volumes" Dec 01 20:15:43 crc kubenswrapper[4960]: I1201 20:15:43.870808 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:15:43 crc kubenswrapper[4960]: I1201 20:15:43.872187 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:15:45 crc kubenswrapper[4960]: I1201 20:15:45.041091 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cloudkitty-storageinit-bsnf5"] Dec 01 20:15:45 crc kubenswrapper[4960]: I1201 20:15:45.051560 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cloudkitty-storageinit-bsnf5"] Dec 01 20:15:45 crc kubenswrapper[4960]: I1201 20:15:45.335614 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c330c49-1896-4917-a0cf-4d5d5f59f25c" path="/var/lib/kubelet/pods/4c330c49-1896-4917-a0cf-4d5d5f59f25c/volumes" Dec 01 20:15:57 crc kubenswrapper[4960]: I1201 20:15:57.763689 4960 generic.go:334] "Generic (PLEG): container finished" podID="4aa6823d-0480-490a-977b-801110a10181" containerID="7735832ac57eda6b9491fbea03c25460054b9be3104cb7201982c4e5c0dcf540" exitCode=0 Dec 01 20:15:57 crc kubenswrapper[4960]: I1201 20:15:57.763754 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" event={"ID":"4aa6823d-0480-490a-977b-801110a10181","Type":"ContainerDied","Data":"7735832ac57eda6b9491fbea03c25460054b9be3104cb7201982c4e5c0dcf540"} Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.338921 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.432991 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory\") pod \"4aa6823d-0480-490a-977b-801110a10181\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.433214 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64t5m\" (UniqueName: \"kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m\") pod \"4aa6823d-0480-490a-977b-801110a10181\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.433377 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key\") pod \"4aa6823d-0480-490a-977b-801110a10181\" (UID: \"4aa6823d-0480-490a-977b-801110a10181\") " Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.472303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m" (OuterVolumeSpecName: "kube-api-access-64t5m") pod "4aa6823d-0480-490a-977b-801110a10181" (UID: "4aa6823d-0480-490a-977b-801110a10181"). InnerVolumeSpecName "kube-api-access-64t5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.515303 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory" (OuterVolumeSpecName: "inventory") pod "4aa6823d-0480-490a-977b-801110a10181" (UID: "4aa6823d-0480-490a-977b-801110a10181"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.536816 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.536865 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64t5m\" (UniqueName: \"kubernetes.io/projected/4aa6823d-0480-490a-977b-801110a10181-kube-api-access-64t5m\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.558885 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4aa6823d-0480-490a-977b-801110a10181" (UID: "4aa6823d-0480-490a-977b-801110a10181"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.638965 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4aa6823d-0480-490a-977b-801110a10181-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.792605 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" event={"ID":"4aa6823d-0480-490a-977b-801110a10181","Type":"ContainerDied","Data":"05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d"} Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.792639 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-9bb2x" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.792643 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05e7d15d82cc29341608460fe83a6619ba8c069c7d79fcddc3041a94d559ff2d" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.898827 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2"] Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899454 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899485 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899509 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899522 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899549 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899561 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="extract-utilities" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899588 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899599 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899616 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899625 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899649 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa6823d-0480-490a-977b-801110a10181" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899661 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa6823d-0480-490a-977b-801110a10181" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899685 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899695 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899717 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899729 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="extract-content" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899742 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899750 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: E1201 20:15:59.899788 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.899814 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.900318 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc3f5b8-f5fb-4913-b85e-b2588a7bde5f" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.900383 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa6823d-0480-490a-977b-801110a10181" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.900409 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="79f3119a-9f60-4683-b5eb-7b6e2ca2dfb6" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.900418 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0043e5b-1732-44f2-8afe-4c5a2e52c130" containerName="registry-server" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.901552 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.905384 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.905987 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.906107 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.910135 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.919216 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2"] Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.945905 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkjhl\" (UniqueName: \"kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.945958 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:15:59 crc kubenswrapper[4960]: I1201 20:15:59.946201 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.048620 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.048778 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkjhl\" (UniqueName: \"kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.048821 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.054244 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.054319 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.069074 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkjhl\" (UniqueName: \"kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.229029 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.846299 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2"] Dec 01 20:16:00 crc kubenswrapper[4960]: I1201 20:16:00.850950 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:16:01 crc kubenswrapper[4960]: I1201 20:16:01.812125 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" event={"ID":"f45d70f4-68fe-491f-aa15-4e7d6a0a7812","Type":"ContainerStarted","Data":"839defe6c3394aaf14f3800a8fa779085cfa2d86ea154865338ced43d3b1d9a9"} Dec 01 20:16:03 crc kubenswrapper[4960]: I1201 20:16:03.835033 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" event={"ID":"f45d70f4-68fe-491f-aa15-4e7d6a0a7812","Type":"ContainerStarted","Data":"932871e869e1bf78a9b8a9cd1c77fba6f9356a55ad5eeb2cdcaff5862d2af9d5"} Dec 01 20:16:03 crc kubenswrapper[4960]: I1201 20:16:03.858971 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" podStartSLOduration=3.059531929 podStartE2EDuration="4.85895095s" podCreationTimestamp="2025-12-01 20:15:59 +0000 UTC" firstStartedPulling="2025-12-01 20:16:00.850655549 +0000 UTC m=+2196.138147218" lastFinishedPulling="2025-12-01 20:16:02.65007454 +0000 UTC m=+2197.937566239" observedRunningTime="2025-12-01 20:16:03.852720415 +0000 UTC m=+2199.140212084" watchObservedRunningTime="2025-12-01 20:16:03.85895095 +0000 UTC m=+2199.146442629" Dec 01 20:16:13 crc kubenswrapper[4960]: I1201 20:16:13.871314 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:16:13 crc kubenswrapper[4960]: I1201 20:16:13.871847 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:16:13 crc kubenswrapper[4960]: I1201 20:16:13.871889 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:16:13 crc kubenswrapper[4960]: I1201 20:16:13.872763 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:16:13 crc kubenswrapper[4960]: I1201 20:16:13.872822 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2" gracePeriod=600 Dec 01 20:16:14 crc kubenswrapper[4960]: I1201 20:16:14.949868 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2" exitCode=0 Dec 01 20:16:14 crc kubenswrapper[4960]: I1201 20:16:14.949944 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2"} Dec 01 20:16:14 crc kubenswrapper[4960]: I1201 20:16:14.950485 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933"} Dec 01 20:16:14 crc kubenswrapper[4960]: I1201 20:16:14.950512 4960 scope.go:117] "RemoveContainer" containerID="7daebf4f2797309473252274fd9a9468c90f04e34bfaa1141b060ae1c7a60d01" Dec 01 20:16:15 crc kubenswrapper[4960]: I1201 20:16:15.447605 4960 scope.go:117] "RemoveContainer" containerID="12ea03fb40e20df4f24a4d5c3267e8540db8e91cc9301298b637e189264ff046" Dec 01 20:16:15 crc kubenswrapper[4960]: I1201 20:16:15.482708 4960 scope.go:117] "RemoveContainer" containerID="da51dfbd41632d77feaf2e9ae1b58baa07d8556b44d96181be1f69390f4bc9c0" Dec 01 20:16:57 crc kubenswrapper[4960]: I1201 20:16:57.401769 4960 generic.go:334] "Generic (PLEG): container finished" podID="f45d70f4-68fe-491f-aa15-4e7d6a0a7812" containerID="932871e869e1bf78a9b8a9cd1c77fba6f9356a55ad5eeb2cdcaff5862d2af9d5" exitCode=0 Dec 01 20:16:57 crc kubenswrapper[4960]: I1201 20:16:57.401847 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" event={"ID":"f45d70f4-68fe-491f-aa15-4e7d6a0a7812","Type":"ContainerDied","Data":"932871e869e1bf78a9b8a9cd1c77fba6f9356a55ad5eeb2cdcaff5862d2af9d5"} Dec 01 20:16:58 crc kubenswrapper[4960]: I1201 20:16:58.894338 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.056233 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory\") pod \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.056398 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key\") pod \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.056452 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkjhl\" (UniqueName: \"kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl\") pod \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\" (UID: \"f45d70f4-68fe-491f-aa15-4e7d6a0a7812\") " Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.062194 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl" (OuterVolumeSpecName: "kube-api-access-pkjhl") pod "f45d70f4-68fe-491f-aa15-4e7d6a0a7812" (UID: "f45d70f4-68fe-491f-aa15-4e7d6a0a7812"). InnerVolumeSpecName "kube-api-access-pkjhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.091367 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f45d70f4-68fe-491f-aa15-4e7d6a0a7812" (UID: "f45d70f4-68fe-491f-aa15-4e7d6a0a7812"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.091740 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory" (OuterVolumeSpecName: "inventory") pod "f45d70f4-68fe-491f-aa15-4e7d6a0a7812" (UID: "f45d70f4-68fe-491f-aa15-4e7d6a0a7812"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.160420 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.160470 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.160480 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkjhl\" (UniqueName: \"kubernetes.io/projected/f45d70f4-68fe-491f-aa15-4e7d6a0a7812-kube-api-access-pkjhl\") on node \"crc\" DevicePath \"\"" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.430476 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" event={"ID":"f45d70f4-68fe-491f-aa15-4e7d6a0a7812","Type":"ContainerDied","Data":"839defe6c3394aaf14f3800a8fa779085cfa2d86ea154865338ced43d3b1d9a9"} Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.430905 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="839defe6c3394aaf14f3800a8fa779085cfa2d86ea154865338ced43d3b1d9a9" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.430566 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.514613 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mmbqh"] Dec 01 20:16:59 crc kubenswrapper[4960]: E1201 20:16:59.518048 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45d70f4-68fe-491f-aa15-4e7d6a0a7812" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.518183 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45d70f4-68fe-491f-aa15-4e7d6a0a7812" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.518471 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45d70f4-68fe-491f-aa15-4e7d6a0a7812" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.519949 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.523850 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.524078 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.524242 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.524451 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.530806 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mmbqh"] Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.671977 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2kzw\" (UniqueName: \"kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.672092 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.672404 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.774642 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2kzw\" (UniqueName: \"kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.774758 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.774857 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.781316 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.782690 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.792691 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2kzw\" (UniqueName: \"kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw\") pod \"ssh-known-hosts-edpm-deployment-mmbqh\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:16:59 crc kubenswrapper[4960]: I1201 20:16:59.839496 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:17:00 crc kubenswrapper[4960]: I1201 20:17:00.399578 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mmbqh"] Dec 01 20:17:00 crc kubenswrapper[4960]: I1201 20:17:00.443858 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" event={"ID":"9ab87b90-6085-4c16-a656-a6775aeb0b68","Type":"ContainerStarted","Data":"314633d8c80a00303d76aeabb9f4270ce9573d389c1e643c5fe6941711bad8ac"} Dec 01 20:17:01 crc kubenswrapper[4960]: I1201 20:17:01.453381 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" event={"ID":"9ab87b90-6085-4c16-a656-a6775aeb0b68","Type":"ContainerStarted","Data":"78431c8cd814e28c40d46b60cd51a08c9f79211e980057ebb9392291d10a2071"} Dec 01 20:17:01 crc kubenswrapper[4960]: I1201 20:17:01.469533 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" podStartSLOduration=1.917242662 podStartE2EDuration="2.469511401s" podCreationTimestamp="2025-12-01 20:16:59 +0000 UTC" firstStartedPulling="2025-12-01 20:17:00.401189977 +0000 UTC m=+2255.688681646" lastFinishedPulling="2025-12-01 20:17:00.953458686 +0000 UTC m=+2256.240950385" observedRunningTime="2025-12-01 20:17:01.466779366 +0000 UTC m=+2256.754271045" watchObservedRunningTime="2025-12-01 20:17:01.469511401 +0000 UTC m=+2256.757003070" Dec 01 20:17:08 crc kubenswrapper[4960]: I1201 20:17:08.546834 4960 generic.go:334] "Generic (PLEG): container finished" podID="9ab87b90-6085-4c16-a656-a6775aeb0b68" containerID="78431c8cd814e28c40d46b60cd51a08c9f79211e980057ebb9392291d10a2071" exitCode=0 Dec 01 20:17:08 crc kubenswrapper[4960]: I1201 20:17:08.546996 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" event={"ID":"9ab87b90-6085-4c16-a656-a6775aeb0b68","Type":"ContainerDied","Data":"78431c8cd814e28c40d46b60cd51a08c9f79211e980057ebb9392291d10a2071"} Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.026710 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.213668 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2kzw\" (UniqueName: \"kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw\") pod \"9ab87b90-6085-4c16-a656-a6775aeb0b68\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.214109 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0\") pod \"9ab87b90-6085-4c16-a656-a6775aeb0b68\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.214195 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam\") pod \"9ab87b90-6085-4c16-a656-a6775aeb0b68\" (UID: \"9ab87b90-6085-4c16-a656-a6775aeb0b68\") " Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.220760 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw" (OuterVolumeSpecName: "kube-api-access-v2kzw") pod "9ab87b90-6085-4c16-a656-a6775aeb0b68" (UID: "9ab87b90-6085-4c16-a656-a6775aeb0b68"). InnerVolumeSpecName "kube-api-access-v2kzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.244822 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9ab87b90-6085-4c16-a656-a6775aeb0b68" (UID: "9ab87b90-6085-4c16-a656-a6775aeb0b68"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.249517 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9ab87b90-6085-4c16-a656-a6775aeb0b68" (UID: "9ab87b90-6085-4c16-a656-a6775aeb0b68"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.316560 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.316598 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2kzw\" (UniqueName: \"kubernetes.io/projected/9ab87b90-6085-4c16-a656-a6775aeb0b68-kube-api-access-v2kzw\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.316609 4960 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ab87b90-6085-4c16-a656-a6775aeb0b68-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.569189 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" event={"ID":"9ab87b90-6085-4c16-a656-a6775aeb0b68","Type":"ContainerDied","Data":"314633d8c80a00303d76aeabb9f4270ce9573d389c1e643c5fe6941711bad8ac"} Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.569236 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="314633d8c80a00303d76aeabb9f4270ce9573d389c1e643c5fe6941711bad8ac" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.569265 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mmbqh" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.643780 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8"] Dec 01 20:17:10 crc kubenswrapper[4960]: E1201 20:17:10.644372 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab87b90-6085-4c16-a656-a6775aeb0b68" containerName="ssh-known-hosts-edpm-deployment" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.644391 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab87b90-6085-4c16-a656-a6775aeb0b68" containerName="ssh-known-hosts-edpm-deployment" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.644625 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab87b90-6085-4c16-a656-a6775aeb0b68" containerName="ssh-known-hosts-edpm-deployment" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.645584 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.648014 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.648352 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.651805 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.656514 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.659240 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8"] Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.723460 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.723865 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmwnx\" (UniqueName: \"kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.724001 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.826497 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.826888 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmwnx\" (UniqueName: \"kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.827005 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.832645 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.832748 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.861643 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmwnx\" (UniqueName: \"kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-t8vj8\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:10 crc kubenswrapper[4960]: I1201 20:17:10.972918 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:11 crc kubenswrapper[4960]: I1201 20:17:11.486932 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8"] Dec 01 20:17:11 crc kubenswrapper[4960]: I1201 20:17:11.578057 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" event={"ID":"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34","Type":"ContainerStarted","Data":"9d8223e44338e78dfe6d136fb0da754f206bc5948f926b4856ed4820376b58d3"} Dec 01 20:17:13 crc kubenswrapper[4960]: I1201 20:17:13.602005 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" event={"ID":"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34","Type":"ContainerStarted","Data":"2e763c31edde21a4eba2eb012d4cb8987d05f6df583a319a7ee75864e39901bc"} Dec 01 20:17:13 crc kubenswrapper[4960]: I1201 20:17:13.632314 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" podStartSLOduration=2.345011419 podStartE2EDuration="3.632295558s" podCreationTimestamp="2025-12-01 20:17:10 +0000 UTC" firstStartedPulling="2025-12-01 20:17:11.491434397 +0000 UTC m=+2266.778926066" lastFinishedPulling="2025-12-01 20:17:12.778718506 +0000 UTC m=+2268.066210205" observedRunningTime="2025-12-01 20:17:13.6259587 +0000 UTC m=+2268.913450379" watchObservedRunningTime="2025-12-01 20:17:13.632295558 +0000 UTC m=+2268.919787227" Dec 01 20:17:22 crc kubenswrapper[4960]: I1201 20:17:22.698523 4960 generic.go:334] "Generic (PLEG): container finished" podID="91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" containerID="2e763c31edde21a4eba2eb012d4cb8987d05f6df583a319a7ee75864e39901bc" exitCode=0 Dec 01 20:17:22 crc kubenswrapper[4960]: I1201 20:17:22.698630 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" event={"ID":"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34","Type":"ContainerDied","Data":"2e763c31edde21a4eba2eb012d4cb8987d05f6df583a319a7ee75864e39901bc"} Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.214632 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.256461 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key\") pod \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.256569 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmwnx\" (UniqueName: \"kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx\") pod \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.256761 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory\") pod \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\" (UID: \"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34\") " Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.269490 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx" (OuterVolumeSpecName: "kube-api-access-fmwnx") pod "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" (UID: "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34"). InnerVolumeSpecName "kube-api-access-fmwnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.298095 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory" (OuterVolumeSpecName: "inventory") pod "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" (UID: "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.304487 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" (UID: "91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.359514 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.359547 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmwnx\" (UniqueName: \"kubernetes.io/projected/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-kube-api-access-fmwnx\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.359557 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.728851 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" event={"ID":"91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34","Type":"ContainerDied","Data":"9d8223e44338e78dfe6d136fb0da754f206bc5948f926b4856ed4820376b58d3"} Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.728898 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d8223e44338e78dfe6d136fb0da754f206bc5948f926b4856ed4820376b58d3" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.728951 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-t8vj8" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.849714 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc"] Dec 01 20:17:24 crc kubenswrapper[4960]: E1201 20:17:24.850744 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.850766 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.851001 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.851903 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.854275 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.854454 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.854860 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.855158 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.864726 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc"] Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.874157 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.874252 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.874412 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvmlg\" (UniqueName: \"kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.977412 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvmlg\" (UniqueName: \"kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.977582 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.977700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.988919 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.989055 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:24 crc kubenswrapper[4960]: I1201 20:17:24.996688 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvmlg\" (UniqueName: \"kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:25 crc kubenswrapper[4960]: I1201 20:17:25.169581 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:25 crc kubenswrapper[4960]: I1201 20:17:25.728359 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc"] Dec 01 20:17:26 crc kubenswrapper[4960]: I1201 20:17:26.758639 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" event={"ID":"1d22d084-b469-42e1-afdf-6c8ba0c4c59e","Type":"ContainerStarted","Data":"accb3fcbccde330afacb4b11bc0d35c344748294158822be000ba5bf59bfc23b"} Dec 01 20:17:27 crc kubenswrapper[4960]: I1201 20:17:27.568029 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:17:28 crc kubenswrapper[4960]: I1201 20:17:28.785617 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" event={"ID":"1d22d084-b469-42e1-afdf-6c8ba0c4c59e","Type":"ContainerStarted","Data":"32970dfa8cba2521b1269eb4dec5af1f1042eb732852a6a97bb2fd83b7d0f416"} Dec 01 20:17:28 crc kubenswrapper[4960]: I1201 20:17:28.804010 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" podStartSLOduration=2.990661614 podStartE2EDuration="4.803990002s" podCreationTimestamp="2025-12-01 20:17:24 +0000 UTC" firstStartedPulling="2025-12-01 20:17:25.752485042 +0000 UTC m=+2281.039976751" lastFinishedPulling="2025-12-01 20:17:27.56581347 +0000 UTC m=+2282.853305139" observedRunningTime="2025-12-01 20:17:28.801798513 +0000 UTC m=+2284.089290212" watchObservedRunningTime="2025-12-01 20:17:28.803990002 +0000 UTC m=+2284.091481661" Dec 01 20:17:37 crc kubenswrapper[4960]: I1201 20:17:37.878341 4960 generic.go:334] "Generic (PLEG): container finished" podID="1d22d084-b469-42e1-afdf-6c8ba0c4c59e" containerID="32970dfa8cba2521b1269eb4dec5af1f1042eb732852a6a97bb2fd83b7d0f416" exitCode=0 Dec 01 20:17:37 crc kubenswrapper[4960]: I1201 20:17:37.878462 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" event={"ID":"1d22d084-b469-42e1-afdf-6c8ba0c4c59e","Type":"ContainerDied","Data":"32970dfa8cba2521b1269eb4dec5af1f1042eb732852a6a97bb2fd83b7d0f416"} Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.353829 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.487642 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvmlg\" (UniqueName: \"kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg\") pod \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.487730 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key\") pod \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.487950 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory\") pod \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\" (UID: \"1d22d084-b469-42e1-afdf-6c8ba0c4c59e\") " Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.494213 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg" (OuterVolumeSpecName: "kube-api-access-gvmlg") pod "1d22d084-b469-42e1-afdf-6c8ba0c4c59e" (UID: "1d22d084-b469-42e1-afdf-6c8ba0c4c59e"). InnerVolumeSpecName "kube-api-access-gvmlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.519546 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory" (OuterVolumeSpecName: "inventory") pod "1d22d084-b469-42e1-afdf-6c8ba0c4c59e" (UID: "1d22d084-b469-42e1-afdf-6c8ba0c4c59e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.522481 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d22d084-b469-42e1-afdf-6c8ba0c4c59e" (UID: "1d22d084-b469-42e1-afdf-6c8ba0c4c59e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.590749 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvmlg\" (UniqueName: \"kubernetes.io/projected/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-kube-api-access-gvmlg\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.590800 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.590812 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d22d084-b469-42e1-afdf-6c8ba0c4c59e-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.897891 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" event={"ID":"1d22d084-b469-42e1-afdf-6c8ba0c4c59e","Type":"ContainerDied","Data":"accb3fcbccde330afacb4b11bc0d35c344748294158822be000ba5bf59bfc23b"} Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.897939 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="accb3fcbccde330afacb4b11bc0d35c344748294158822be000ba5bf59bfc23b" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.897975 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.980560 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr"] Dec 01 20:17:39 crc kubenswrapper[4960]: E1201 20:17:39.981397 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d22d084-b469-42e1-afdf-6c8ba0c4c59e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.981417 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d22d084-b469-42e1-afdf-6c8ba0c4c59e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.981651 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d22d084-b469-42e1-afdf-6c8ba0c4c59e" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.982528 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.986904 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987054 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987244 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987395 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987533 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987548 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987649 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.987712 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 01 20:17:39 crc kubenswrapper[4960]: I1201 20:17:39.992974 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr"] Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106381 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106452 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106483 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlfq\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106548 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106576 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106612 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106656 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106941 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.106990 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.107157 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.107237 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.107298 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.107347 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210199 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210295 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210325 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlfq\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210421 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210456 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210500 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210541 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210585 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210625 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210654 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210734 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210761 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.210792 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.215305 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.215761 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.216285 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.216308 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.216291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.216407 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.216755 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.217044 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.217206 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.217312 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.218092 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.218376 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.218205 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.232591 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlfq\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.312300 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.856574 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr"] Dec 01 20:17:40 crc kubenswrapper[4960]: I1201 20:17:40.912024 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" event={"ID":"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71","Type":"ContainerStarted","Data":"119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11"} Dec 01 20:17:41 crc kubenswrapper[4960]: I1201 20:17:41.924102 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" event={"ID":"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71","Type":"ContainerStarted","Data":"cb132c8dfb3609007f48b157fb8580897cbe6079fe1e98cb80eb7a00b5e9256a"} Dec 01 20:17:41 crc kubenswrapper[4960]: I1201 20:17:41.955970 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" podStartSLOduration=2.212329256 podStartE2EDuration="2.955947055s" podCreationTimestamp="2025-12-01 20:17:39 +0000 UTC" firstStartedPulling="2025-12-01 20:17:40.860942315 +0000 UTC m=+2296.148433984" lastFinishedPulling="2025-12-01 20:17:41.604560114 +0000 UTC m=+2296.892051783" observedRunningTime="2025-12-01 20:17:41.946426127 +0000 UTC m=+2297.233917796" watchObservedRunningTime="2025-12-01 20:17:41.955947055 +0000 UTC m=+2297.243438734" Dec 01 20:18:23 crc kubenswrapper[4960]: I1201 20:18:23.530539 4960 generic.go:334] "Generic (PLEG): container finished" podID="cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" containerID="cb132c8dfb3609007f48b157fb8580897cbe6079fe1e98cb80eb7a00b5e9256a" exitCode=0 Dec 01 20:18:23 crc kubenswrapper[4960]: I1201 20:18:23.530641 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" event={"ID":"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71","Type":"ContainerDied","Data":"cb132c8dfb3609007f48b157fb8580897cbe6079fe1e98cb80eb7a00b5e9256a"} Dec 01 20:18:24 crc kubenswrapper[4960]: I1201 20:18:24.987037 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045696 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045779 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045826 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045863 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045891 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.045925 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046761 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlfq\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046786 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046805 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046912 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046946 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.046965 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.047029 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0\") pod \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\" (UID: \"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71\") " Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.053346 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.053391 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.054328 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.054549 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq" (OuterVolumeSpecName: "kube-api-access-tqlfq") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "kube-api-access-tqlfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.055256 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.056272 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.059005 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.059366 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.059678 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.060253 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.064651 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.080957 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.083573 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.084971 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory" (OuterVolumeSpecName: "inventory") pod "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" (UID: "cc7c39de-f6ea-4f00-ae86-a8f74ec35b71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149904 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149942 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149959 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149971 4960 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149983 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.149995 4960 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150006 4960 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150017 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150028 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlfq\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-kube-api-access-tqlfq\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150038 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150048 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150058 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150071 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.150083 4960 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc7c39de-f6ea-4f00-ae86-a8f74ec35b71-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.552972 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" event={"ID":"cc7c39de-f6ea-4f00-ae86-a8f74ec35b71","Type":"ContainerDied","Data":"119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11"} Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.553619 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.553572 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr" Dec 01 20:18:25 crc kubenswrapper[4960]: E1201 20:18:25.592938 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache]" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.659189 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97"] Dec 01 20:18:25 crc kubenswrapper[4960]: E1201 20:18:25.659705 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.659724 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.659921 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7c39de-f6ea-4f00-ae86-a8f74ec35b71" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.660739 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.665604 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.665794 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.665897 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.667431 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.667642 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.679280 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97"] Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.763378 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5kdz\" (UniqueName: \"kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.763507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.763627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.763675 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.763760 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.866306 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.866375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.866428 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.866466 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5kdz\" (UniqueName: \"kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.866522 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.867427 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.872328 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.872504 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.872605 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.900372 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5kdz\" (UniqueName: \"kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-bhl97\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:25 crc kubenswrapper[4960]: I1201 20:18:25.989836 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:18:26 crc kubenswrapper[4960]: I1201 20:18:26.534069 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97"] Dec 01 20:18:26 crc kubenswrapper[4960]: I1201 20:18:26.564014 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" event={"ID":"95e8c909-a7d7-4461-8fc7-c7644ad41000","Type":"ContainerStarted","Data":"b3cf9285ba6c7780fbb0cc214b4bc1a55dd65ed2362ec104e6f59eae5e5e7e43"} Dec 01 20:18:28 crc kubenswrapper[4960]: I1201 20:18:28.587816 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" event={"ID":"95e8c909-a7d7-4461-8fc7-c7644ad41000","Type":"ContainerStarted","Data":"f4fde6abc5b8a6d86e8359bfa9611d66e2851bfdea785fbfc8182c52810a3c2a"} Dec 01 20:18:28 crc kubenswrapper[4960]: I1201 20:18:28.605316 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" podStartSLOduration=2.650875211 podStartE2EDuration="3.605298421s" podCreationTimestamp="2025-12-01 20:18:25 +0000 UTC" firstStartedPulling="2025-12-01 20:18:26.534369815 +0000 UTC m=+2341.821861494" lastFinishedPulling="2025-12-01 20:18:27.488793005 +0000 UTC m=+2342.776284704" observedRunningTime="2025-12-01 20:18:28.604123394 +0000 UTC m=+2343.891615073" watchObservedRunningTime="2025-12-01 20:18:28.605298421 +0000 UTC m=+2343.892790090" Dec 01 20:18:35 crc kubenswrapper[4960]: E1201 20:18:35.873472 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache]" Dec 01 20:18:43 crc kubenswrapper[4960]: I1201 20:18:43.870617 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:18:43 crc kubenswrapper[4960]: I1201 20:18:43.871097 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:18:46 crc kubenswrapper[4960]: E1201 20:18:46.145245 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache]" Dec 01 20:18:56 crc kubenswrapper[4960]: E1201 20:18:56.447429 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache]" Dec 01 20:19:06 crc kubenswrapper[4960]: E1201 20:19:06.718492 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache]" Dec 01 20:19:13 crc kubenswrapper[4960]: I1201 20:19:13.870426 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:19:13 crc kubenswrapper[4960]: I1201 20:19:13.872339 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:19:16 crc kubenswrapper[4960]: E1201 20:19:16.983085 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7c39de_f6ea_4f00_ae86_a8f74ec35b71.slice/crio-119033125acdc4db84cd5bf4350915be25fa2bbeb5eb59711205c6ff98c1bc11\": RecentStats: unable to find data in memory cache]" Dec 01 20:19:37 crc kubenswrapper[4960]: E1201 20:19:37.529440 4960 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95e8c909_a7d7_4461_8fc7_c7644ad41000.slice/crio-conmon-f4fde6abc5b8a6d86e8359bfa9611d66e2851bfdea785fbfc8182c52810a3c2a.scope\": RecentStats: unable to find data in memory cache]" Dec 01 20:19:38 crc kubenswrapper[4960]: I1201 20:19:38.314847 4960 generic.go:334] "Generic (PLEG): container finished" podID="95e8c909-a7d7-4461-8fc7-c7644ad41000" containerID="f4fde6abc5b8a6d86e8359bfa9611d66e2851bfdea785fbfc8182c52810a3c2a" exitCode=0 Dec 01 20:19:38 crc kubenswrapper[4960]: I1201 20:19:38.314903 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" event={"ID":"95e8c909-a7d7-4461-8fc7-c7644ad41000","Type":"ContainerDied","Data":"f4fde6abc5b8a6d86e8359bfa9611d66e2851bfdea785fbfc8182c52810a3c2a"} Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.787231 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.888405 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0\") pod \"95e8c909-a7d7-4461-8fc7-c7644ad41000\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.888476 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle\") pod \"95e8c909-a7d7-4461-8fc7-c7644ad41000\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.888598 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory\") pod \"95e8c909-a7d7-4461-8fc7-c7644ad41000\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.888723 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key\") pod \"95e8c909-a7d7-4461-8fc7-c7644ad41000\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.889059 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5kdz\" (UniqueName: \"kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz\") pod \"95e8c909-a7d7-4461-8fc7-c7644ad41000\" (UID: \"95e8c909-a7d7-4461-8fc7-c7644ad41000\") " Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.895721 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz" (OuterVolumeSpecName: "kube-api-access-b5kdz") pod "95e8c909-a7d7-4461-8fc7-c7644ad41000" (UID: "95e8c909-a7d7-4461-8fc7-c7644ad41000"). InnerVolumeSpecName "kube-api-access-b5kdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.905701 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "95e8c909-a7d7-4461-8fc7-c7644ad41000" (UID: "95e8c909-a7d7-4461-8fc7-c7644ad41000"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.916463 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory" (OuterVolumeSpecName: "inventory") pod "95e8c909-a7d7-4461-8fc7-c7644ad41000" (UID: "95e8c909-a7d7-4461-8fc7-c7644ad41000"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.923319 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "95e8c909-a7d7-4461-8fc7-c7644ad41000" (UID: "95e8c909-a7d7-4461-8fc7-c7644ad41000"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.938220 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "95e8c909-a7d7-4461-8fc7-c7644ad41000" (UID: "95e8c909-a7d7-4461-8fc7-c7644ad41000"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.992495 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.992534 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5kdz\" (UniqueName: \"kubernetes.io/projected/95e8c909-a7d7-4461-8fc7-c7644ad41000-kube-api-access-b5kdz\") on node \"crc\" DevicePath \"\"" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.992548 4960 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.992563 4960 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:19:39 crc kubenswrapper[4960]: I1201 20:19:39.992578 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e8c909-a7d7-4461-8fc7-c7644ad41000-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.333811 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" event={"ID":"95e8c909-a7d7-4461-8fc7-c7644ad41000","Type":"ContainerDied","Data":"b3cf9285ba6c7780fbb0cc214b4bc1a55dd65ed2362ec104e6f59eae5e5e7e43"} Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.333850 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3cf9285ba6c7780fbb0cc214b4bc1a55dd65ed2362ec104e6f59eae5e5e7e43" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.333911 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-bhl97" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.430011 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58"] Dec 01 20:19:40 crc kubenswrapper[4960]: E1201 20:19:40.430686 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e8c909-a7d7-4461-8fc7-c7644ad41000" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.430702 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e8c909-a7d7-4461-8fc7-c7644ad41000" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.430891 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e8c909-a7d7-4461-8fc7-c7644ad41000" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.431596 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.435460 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.436668 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.436746 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.436895 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.436681 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.436993 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.464851 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58"] Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.502575 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l2mh\" (UniqueName: \"kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.502644 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.502698 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.502769 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.502901 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.503079 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.604907 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l2mh\" (UniqueName: \"kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.604986 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.605030 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.605083 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.605236 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.605340 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.609166 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.609350 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.609928 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.610261 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.610331 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.624734 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l2mh\" (UniqueName: \"kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:40 crc kubenswrapper[4960]: I1201 20:19:40.756921 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:19:41 crc kubenswrapper[4960]: I1201 20:19:41.297766 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58"] Dec 01 20:19:41 crc kubenswrapper[4960]: I1201 20:19:41.342148 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" event={"ID":"6e32c90f-f0de-404d-be43-6953814aa12a","Type":"ContainerStarted","Data":"549d94f26b9e19a34ae37eda0e7b491e72f8bcfc34b732a332198b571db94ec8"} Dec 01 20:19:42 crc kubenswrapper[4960]: I1201 20:19:42.354352 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" event={"ID":"6e32c90f-f0de-404d-be43-6953814aa12a","Type":"ContainerStarted","Data":"83131de3b592b19674d7820d9da3d9403927e67b6bbe5a8e12b84f379ff38921"} Dec 01 20:19:42 crc kubenswrapper[4960]: I1201 20:19:42.372647 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" podStartSLOduration=1.820665071 podStartE2EDuration="2.372632085s" podCreationTimestamp="2025-12-01 20:19:40 +0000 UTC" firstStartedPulling="2025-12-01 20:19:41.302357304 +0000 UTC m=+2416.589848983" lastFinishedPulling="2025-12-01 20:19:41.854324328 +0000 UTC m=+2417.141815997" observedRunningTime="2025-12-01 20:19:42.370696635 +0000 UTC m=+2417.658188304" watchObservedRunningTime="2025-12-01 20:19:42.372632085 +0000 UTC m=+2417.660123754" Dec 01 20:19:43 crc kubenswrapper[4960]: I1201 20:19:43.871201 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:19:43 crc kubenswrapper[4960]: I1201 20:19:43.871548 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:19:43 crc kubenswrapper[4960]: I1201 20:19:43.871611 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:19:43 crc kubenswrapper[4960]: I1201 20:19:43.872585 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:19:43 crc kubenswrapper[4960]: I1201 20:19:43.872661 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" gracePeriod=600 Dec 01 20:19:43 crc kubenswrapper[4960]: E1201 20:19:43.993245 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:19:44 crc kubenswrapper[4960]: I1201 20:19:44.375623 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" exitCode=0 Dec 01 20:19:44 crc kubenswrapper[4960]: I1201 20:19:44.375662 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933"} Dec 01 20:19:44 crc kubenswrapper[4960]: I1201 20:19:44.375695 4960 scope.go:117] "RemoveContainer" containerID="96ee832983a8c2f13bfdc07a1c8386729517739e84d2498669406e5cbfab0fc2" Dec 01 20:19:44 crc kubenswrapper[4960]: I1201 20:19:44.376558 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:19:44 crc kubenswrapper[4960]: E1201 20:19:44.376938 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:19:56 crc kubenswrapper[4960]: I1201 20:19:56.324523 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:19:56 crc kubenswrapper[4960]: E1201 20:19:56.325291 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:20:08 crc kubenswrapper[4960]: I1201 20:20:08.324695 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:20:08 crc kubenswrapper[4960]: E1201 20:20:08.325566 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:20:19 crc kubenswrapper[4960]: I1201 20:20:19.325259 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:20:19 crc kubenswrapper[4960]: E1201 20:20:19.326011 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:20:32 crc kubenswrapper[4960]: I1201 20:20:32.324625 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:20:32 crc kubenswrapper[4960]: E1201 20:20:32.325517 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:20:34 crc kubenswrapper[4960]: I1201 20:20:34.922562 4960 generic.go:334] "Generic (PLEG): container finished" podID="6e32c90f-f0de-404d-be43-6953814aa12a" containerID="83131de3b592b19674d7820d9da3d9403927e67b6bbe5a8e12b84f379ff38921" exitCode=0 Dec 01 20:20:34 crc kubenswrapper[4960]: I1201 20:20:34.922647 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" event={"ID":"6e32c90f-f0de-404d-be43-6953814aa12a","Type":"ContainerDied","Data":"83131de3b592b19674d7820d9da3d9403927e67b6bbe5a8e12b84f379ff38921"} Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.394276 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.441554 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.441766 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.441847 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.441938 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.442009 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.442138 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l2mh\" (UniqueName: \"kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh\") pod \"6e32c90f-f0de-404d-be43-6953814aa12a\" (UID: \"6e32c90f-f0de-404d-be43-6953814aa12a\") " Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.449008 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.449007 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh" (OuterVolumeSpecName: "kube-api-access-9l2mh") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "kube-api-access-9l2mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.479140 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.483595 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.487227 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.501076 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory" (OuterVolumeSpecName: "inventory") pod "6e32c90f-f0de-404d-be43-6953814aa12a" (UID: "6e32c90f-f0de-404d-be43-6953814aa12a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545033 4960 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545078 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545093 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l2mh\" (UniqueName: \"kubernetes.io/projected/6e32c90f-f0de-404d-be43-6953814aa12a-kube-api-access-9l2mh\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545106 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545135 4960 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.545147 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e32c90f-f0de-404d-be43-6953814aa12a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.950240 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" event={"ID":"6e32c90f-f0de-404d-be43-6953814aa12a","Type":"ContainerDied","Data":"549d94f26b9e19a34ae37eda0e7b491e72f8bcfc34b732a332198b571db94ec8"} Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.950285 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="549d94f26b9e19a34ae37eda0e7b491e72f8bcfc34b732a332198b571db94ec8" Dec 01 20:20:36 crc kubenswrapper[4960]: I1201 20:20:36.950345 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.057215 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q"] Dec 01 20:20:37 crc kubenswrapper[4960]: E1201 20:20:37.057617 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e32c90f-f0de-404d-be43-6953814aa12a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.057633 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e32c90f-f0de-404d-be43-6953814aa12a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.057910 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e32c90f-f0de-404d-be43-6953814aa12a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.058648 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.060901 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.061080 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.064607 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.066985 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.067188 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.086259 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q"] Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.154600 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.154654 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.154820 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.155023 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdwnq\" (UniqueName: \"kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.155106 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.256698 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.256810 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdwnq\" (UniqueName: \"kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.256853 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.257840 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.257923 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.261762 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.261887 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.262103 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.262427 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.288640 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdwnq\" (UniqueName: \"kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.375633 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:20:37 crc kubenswrapper[4960]: I1201 20:20:37.951514 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q"] Dec 01 20:20:38 crc kubenswrapper[4960]: I1201 20:20:38.987854 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" event={"ID":"4c5eadde-8ce4-4009-9128-66d4d9657dfc","Type":"ContainerStarted","Data":"b3a607404d60c7f33d6ede58738937ab8f824766f9a965c51d9b399d6901df58"} Dec 01 20:20:40 crc kubenswrapper[4960]: I1201 20:20:40.002570 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" event={"ID":"4c5eadde-8ce4-4009-9128-66d4d9657dfc","Type":"ContainerStarted","Data":"3e7c986adbed560963c6a6de0cd2bed290e72014cbe40747e401d52217fe0fcb"} Dec 01 20:20:40 crc kubenswrapper[4960]: I1201 20:20:40.032564 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" podStartSLOduration=1.911907491 podStartE2EDuration="3.032547818s" podCreationTimestamp="2025-12-01 20:20:37 +0000 UTC" firstStartedPulling="2025-12-01 20:20:37.95836919 +0000 UTC m=+2473.245860859" lastFinishedPulling="2025-12-01 20:20:39.079009517 +0000 UTC m=+2474.366501186" observedRunningTime="2025-12-01 20:20:40.023607037 +0000 UTC m=+2475.311098706" watchObservedRunningTime="2025-12-01 20:20:40.032547818 +0000 UTC m=+2475.320039487" Dec 01 20:20:47 crc kubenswrapper[4960]: I1201 20:20:47.324803 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:20:47 crc kubenswrapper[4960]: E1201 20:20:47.325801 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:20:58 crc kubenswrapper[4960]: I1201 20:20:58.324672 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:20:58 crc kubenswrapper[4960]: E1201 20:20:58.325408 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:21:09 crc kubenswrapper[4960]: I1201 20:21:09.324975 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:21:09 crc kubenswrapper[4960]: E1201 20:21:09.325981 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:21:21 crc kubenswrapper[4960]: I1201 20:21:21.324839 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:21:21 crc kubenswrapper[4960]: E1201 20:21:21.325918 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:21:32 crc kubenswrapper[4960]: I1201 20:21:32.324088 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:21:32 crc kubenswrapper[4960]: E1201 20:21:32.325006 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:21:43 crc kubenswrapper[4960]: I1201 20:21:43.324604 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:21:43 crc kubenswrapper[4960]: E1201 20:21:43.325985 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:21:56 crc kubenswrapper[4960]: I1201 20:21:56.324847 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:21:56 crc kubenswrapper[4960]: E1201 20:21:56.325977 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:22:08 crc kubenswrapper[4960]: I1201 20:22:08.324610 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:22:08 crc kubenswrapper[4960]: E1201 20:22:08.325465 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:22:21 crc kubenswrapper[4960]: I1201 20:22:21.324390 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:22:21 crc kubenswrapper[4960]: E1201 20:22:21.325453 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:22:32 crc kubenswrapper[4960]: I1201 20:22:32.325007 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:22:32 crc kubenswrapper[4960]: E1201 20:22:32.325999 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:22:45 crc kubenswrapper[4960]: I1201 20:22:45.329968 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:22:45 crc kubenswrapper[4960]: E1201 20:22:45.330955 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:22:59 crc kubenswrapper[4960]: I1201 20:22:59.324617 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:22:59 crc kubenswrapper[4960]: E1201 20:22:59.325433 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:23:11 crc kubenswrapper[4960]: I1201 20:23:11.324221 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:23:11 crc kubenswrapper[4960]: E1201 20:23:11.325284 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:23:23 crc kubenswrapper[4960]: I1201 20:23:23.323888 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:23:23 crc kubenswrapper[4960]: E1201 20:23:23.324879 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:23:35 crc kubenswrapper[4960]: I1201 20:23:35.339677 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:23:35 crc kubenswrapper[4960]: E1201 20:23:35.341392 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:23:48 crc kubenswrapper[4960]: I1201 20:23:48.325199 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:23:48 crc kubenswrapper[4960]: E1201 20:23:48.326244 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:24:02 crc kubenswrapper[4960]: I1201 20:24:02.325163 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:24:02 crc kubenswrapper[4960]: E1201 20:24:02.326092 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:24:15 crc kubenswrapper[4960]: I1201 20:24:15.332951 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:24:15 crc kubenswrapper[4960]: E1201 20:24:15.333718 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.844284 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.848325 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.876887 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.981411 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.982098 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:22 crc kubenswrapper[4960]: I1201 20:24:22.982236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824z4\" (UniqueName: \"kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.084265 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.084343 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824z4\" (UniqueName: \"kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.084417 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.084866 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.084879 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.107727 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824z4\" (UniqueName: \"kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4\") pod \"redhat-marketplace-lm26m\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.186504 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:23 crc kubenswrapper[4960]: I1201 20:24:23.773561 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:24 crc kubenswrapper[4960]: I1201 20:24:24.487660 4960 generic.go:334] "Generic (PLEG): container finished" podID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerID="82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6" exitCode=0 Dec 01 20:24:24 crc kubenswrapper[4960]: I1201 20:24:24.487798 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerDied","Data":"82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6"} Dec 01 20:24:24 crc kubenswrapper[4960]: I1201 20:24:24.487955 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerStarted","Data":"75e26ea69f6a1054f95111371318af1d09efc2ca0a78f4e3bef03f86a95279e2"} Dec 01 20:24:24 crc kubenswrapper[4960]: I1201 20:24:24.490553 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:24:30 crc kubenswrapper[4960]: I1201 20:24:30.324154 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:24:30 crc kubenswrapper[4960]: E1201 20:24:30.325844 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:24:30 crc kubenswrapper[4960]: I1201 20:24:30.558920 4960 generic.go:334] "Generic (PLEG): container finished" podID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerID="d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef" exitCode=0 Dec 01 20:24:30 crc kubenswrapper[4960]: I1201 20:24:30.558964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerDied","Data":"d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef"} Dec 01 20:24:31 crc kubenswrapper[4960]: I1201 20:24:31.573105 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerStarted","Data":"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048"} Dec 01 20:24:31 crc kubenswrapper[4960]: I1201 20:24:31.596918 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lm26m" podStartSLOduration=3.070274398 podStartE2EDuration="9.596891909s" podCreationTimestamp="2025-12-01 20:24:22 +0000 UTC" firstStartedPulling="2025-12-01 20:24:24.490283718 +0000 UTC m=+2699.777775387" lastFinishedPulling="2025-12-01 20:24:31.016901219 +0000 UTC m=+2706.304392898" observedRunningTime="2025-12-01 20:24:31.587451475 +0000 UTC m=+2706.874943164" watchObservedRunningTime="2025-12-01 20:24:31.596891909 +0000 UTC m=+2706.884383578" Dec 01 20:24:33 crc kubenswrapper[4960]: I1201 20:24:33.187190 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:33 crc kubenswrapper[4960]: I1201 20:24:33.187558 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:33 crc kubenswrapper[4960]: I1201 20:24:33.240596 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:43 crc kubenswrapper[4960]: I1201 20:24:43.238364 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:43 crc kubenswrapper[4960]: I1201 20:24:43.285430 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:43 crc kubenswrapper[4960]: I1201 20:24:43.678063 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lm26m" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="registry-server" containerID="cri-o://4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048" gracePeriod=2 Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.148570 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.324036 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.335925 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824z4\" (UniqueName: \"kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4\") pod \"12e6eb5a-1b3c-482f-95b8-068f33d52425\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.336045 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content\") pod \"12e6eb5a-1b3c-482f-95b8-068f33d52425\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.336070 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities\") pod \"12e6eb5a-1b3c-482f-95b8-068f33d52425\" (UID: \"12e6eb5a-1b3c-482f-95b8-068f33d52425\") " Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.336935 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities" (OuterVolumeSpecName: "utilities") pod "12e6eb5a-1b3c-482f-95b8-068f33d52425" (UID: "12e6eb5a-1b3c-482f-95b8-068f33d52425"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.342584 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4" (OuterVolumeSpecName: "kube-api-access-824z4") pod "12e6eb5a-1b3c-482f-95b8-068f33d52425" (UID: "12e6eb5a-1b3c-482f-95b8-068f33d52425"). InnerVolumeSpecName "kube-api-access-824z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.361290 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12e6eb5a-1b3c-482f-95b8-068f33d52425" (UID: "12e6eb5a-1b3c-482f-95b8-068f33d52425"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.438074 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824z4\" (UniqueName: \"kubernetes.io/projected/12e6eb5a-1b3c-482f-95b8-068f33d52425-kube-api-access-824z4\") on node \"crc\" DevicePath \"\"" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.438131 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.438146 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e6eb5a-1b3c-482f-95b8-068f33d52425-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.689140 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5"} Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.693987 4960 generic.go:334] "Generic (PLEG): container finished" podID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerID="4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048" exitCode=0 Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.694017 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lm26m" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.694036 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerDied","Data":"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048"} Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.694064 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lm26m" event={"ID":"12e6eb5a-1b3c-482f-95b8-068f33d52425","Type":"ContainerDied","Data":"75e26ea69f6a1054f95111371318af1d09efc2ca0a78f4e3bef03f86a95279e2"} Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.694083 4960 scope.go:117] "RemoveContainer" containerID="4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.717308 4960 scope.go:117] "RemoveContainer" containerID="d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.746046 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.754988 4960 scope.go:117] "RemoveContainer" containerID="82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.758461 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lm26m"] Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.803230 4960 scope.go:117] "RemoveContainer" containerID="4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048" Dec 01 20:24:44 crc kubenswrapper[4960]: E1201 20:24:44.804576 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048\": container with ID starting with 4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048 not found: ID does not exist" containerID="4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.804615 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048"} err="failed to get container status \"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048\": rpc error: code = NotFound desc = could not find container \"4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048\": container with ID starting with 4a30874655a33967b121b53df5fc727d57a5aeb80bc7e4a159ca74cc6efc7048 not found: ID does not exist" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.804645 4960 scope.go:117] "RemoveContainer" containerID="d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef" Dec 01 20:24:44 crc kubenswrapper[4960]: E1201 20:24:44.804941 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef\": container with ID starting with d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef not found: ID does not exist" containerID="d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.804963 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef"} err="failed to get container status \"d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef\": rpc error: code = NotFound desc = could not find container \"d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef\": container with ID starting with d9bea75771e4be0198c098bb7363988fe396ef48184f6fbc039210558083b0ef not found: ID does not exist" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.804975 4960 scope.go:117] "RemoveContainer" containerID="82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6" Dec 01 20:24:44 crc kubenswrapper[4960]: E1201 20:24:44.805209 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6\": container with ID starting with 82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6 not found: ID does not exist" containerID="82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6" Dec 01 20:24:44 crc kubenswrapper[4960]: I1201 20:24:44.805230 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6"} err="failed to get container status \"82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6\": rpc error: code = NotFound desc = could not find container \"82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6\": container with ID starting with 82bc41d3983ae083bf999c43ce21974537cf9bbf112880d52698d05911d879c6 not found: ID does not exist" Dec 01 20:24:45 crc kubenswrapper[4960]: I1201 20:24:45.336332 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" path="/var/lib/kubelet/pods/12e6eb5a-1b3c-482f-95b8-068f33d52425/volumes" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.368287 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:22 crc kubenswrapper[4960]: E1201 20:25:22.374775 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="extract-content" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.374819 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="extract-content" Dec 01 20:25:22 crc kubenswrapper[4960]: E1201 20:25:22.374848 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="registry-server" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.374858 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="registry-server" Dec 01 20:25:22 crc kubenswrapper[4960]: E1201 20:25:22.374881 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="extract-utilities" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.374889 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="extract-utilities" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.375317 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e6eb5a-1b3c-482f-95b8-068f33d52425" containerName="registry-server" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.377321 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.387400 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.557930 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.558001 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.558048 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkckb\" (UniqueName: \"kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.659563 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.659651 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkckb\" (UniqueName: \"kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.659851 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.660080 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.660356 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.679068 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkckb\" (UniqueName: \"kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb\") pod \"redhat-operators-8ncft\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:22 crc kubenswrapper[4960]: I1201 20:25:22.744589 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:23 crc kubenswrapper[4960]: I1201 20:25:23.205280 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:24 crc kubenswrapper[4960]: I1201 20:25:24.116708 4960 generic.go:334] "Generic (PLEG): container finished" podID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerID="3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514" exitCode=0 Dec 01 20:25:24 crc kubenswrapper[4960]: I1201 20:25:24.117345 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerDied","Data":"3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514"} Dec 01 20:25:24 crc kubenswrapper[4960]: I1201 20:25:24.117417 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerStarted","Data":"f3eafc6df64776146af160735cab2dde6ca9a6752078f014e1b18576ff92efc2"} Dec 01 20:25:26 crc kubenswrapper[4960]: I1201 20:25:26.142339 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerStarted","Data":"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f"} Dec 01 20:25:29 crc kubenswrapper[4960]: I1201 20:25:29.175485 4960 generic.go:334] "Generic (PLEG): container finished" podID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerID="435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f" exitCode=0 Dec 01 20:25:29 crc kubenswrapper[4960]: I1201 20:25:29.175560 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerDied","Data":"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f"} Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.211296 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerStarted","Data":"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d"} Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.246427 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8ncft" podStartSLOduration=3.080737765 podStartE2EDuration="10.246402814s" podCreationTimestamp="2025-12-01 20:25:22 +0000 UTC" firstStartedPulling="2025-12-01 20:25:24.11973049 +0000 UTC m=+2759.407222179" lastFinishedPulling="2025-12-01 20:25:31.285395549 +0000 UTC m=+2766.572887228" observedRunningTime="2025-12-01 20:25:32.237355652 +0000 UTC m=+2767.524847331" watchObservedRunningTime="2025-12-01 20:25:32.246402814 +0000 UTC m=+2767.533894503" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.608469 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.611226 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.619004 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.667195 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.667282 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd65w\" (UniqueName: \"kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.667348 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.745522 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.745586 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.770128 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.770255 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.770326 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd65w\" (UniqueName: \"kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.770764 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.771098 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.793585 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd65w\" (UniqueName: \"kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w\") pod \"certified-operators-74rqb\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:32 crc kubenswrapper[4960]: I1201 20:25:32.931797 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:33 crc kubenswrapper[4960]: I1201 20:25:33.452218 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:33 crc kubenswrapper[4960]: I1201 20:25:33.793401 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8ncft" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" probeResult="failure" output=< Dec 01 20:25:33 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:25:33 crc kubenswrapper[4960]: > Dec 01 20:25:34 crc kubenswrapper[4960]: I1201 20:25:34.267633 4960 generic.go:334] "Generic (PLEG): container finished" podID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerID="fad8d58c5b7385db018cde92e5cadac22a356decdb69ccaef48f1d5147e75713" exitCode=0 Dec 01 20:25:34 crc kubenswrapper[4960]: I1201 20:25:34.267697 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerDied","Data":"fad8d58c5b7385db018cde92e5cadac22a356decdb69ccaef48f1d5147e75713"} Dec 01 20:25:34 crc kubenswrapper[4960]: I1201 20:25:34.268210 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerStarted","Data":"e156d3e9af0f33f72e15ebb6078314c466da2d962209abdac37b3b5ab56aa7b5"} Dec 01 20:25:36 crc kubenswrapper[4960]: I1201 20:25:36.286555 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerStarted","Data":"0b6382163d0f89c3fca50c66c764e091c8e91f7cc2f3d8e1eb8d7bf5c78022c8"} Dec 01 20:25:37 crc kubenswrapper[4960]: I1201 20:25:37.298517 4960 generic.go:334] "Generic (PLEG): container finished" podID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerID="0b6382163d0f89c3fca50c66c764e091c8e91f7cc2f3d8e1eb8d7bf5c78022c8" exitCode=0 Dec 01 20:25:37 crc kubenswrapper[4960]: I1201 20:25:37.298570 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerDied","Data":"0b6382163d0f89c3fca50c66c764e091c8e91f7cc2f3d8e1eb8d7bf5c78022c8"} Dec 01 20:25:40 crc kubenswrapper[4960]: I1201 20:25:40.334499 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerStarted","Data":"d3772d2f43f340e8ea4471c73c5c4f101479b92fdecc61a9abaef0f3ce9035c1"} Dec 01 20:25:40 crc kubenswrapper[4960]: I1201 20:25:40.361678 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-74rqb" podStartSLOduration=3.371772343 podStartE2EDuration="8.361656902s" podCreationTimestamp="2025-12-01 20:25:32 +0000 UTC" firstStartedPulling="2025-12-01 20:25:34.271212794 +0000 UTC m=+2769.558704463" lastFinishedPulling="2025-12-01 20:25:39.261097333 +0000 UTC m=+2774.548589022" observedRunningTime="2025-12-01 20:25:40.360451735 +0000 UTC m=+2775.647943404" watchObservedRunningTime="2025-12-01 20:25:40.361656902 +0000 UTC m=+2775.649148571" Dec 01 20:25:42 crc kubenswrapper[4960]: I1201 20:25:42.932651 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:42 crc kubenswrapper[4960]: I1201 20:25:42.932963 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:43 crc kubenswrapper[4960]: I1201 20:25:43.804626 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8ncft" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" probeResult="failure" output=< Dec 01 20:25:43 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:25:43 crc kubenswrapper[4960]: > Dec 01 20:25:43 crc kubenswrapper[4960]: I1201 20:25:43.977275 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-74rqb" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="registry-server" probeResult="failure" output=< Dec 01 20:25:43 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:25:43 crc kubenswrapper[4960]: > Dec 01 20:25:49 crc kubenswrapper[4960]: I1201 20:25:49.430269 4960 generic.go:334] "Generic (PLEG): container finished" podID="4c5eadde-8ce4-4009-9128-66d4d9657dfc" containerID="3e7c986adbed560963c6a6de0cd2bed290e72014cbe40747e401d52217fe0fcb" exitCode=0 Dec 01 20:25:49 crc kubenswrapper[4960]: I1201 20:25:49.430362 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" event={"ID":"4c5eadde-8ce4-4009-9128-66d4d9657dfc","Type":"ContainerDied","Data":"3e7c986adbed560963c6a6de0cd2bed290e72014cbe40747e401d52217fe0fcb"} Dec 01 20:25:50 crc kubenswrapper[4960]: I1201 20:25:50.920130 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.057512 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key\") pod \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.057605 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle\") pod \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.057652 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0\") pod \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.057700 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory\") pod \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.057798 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdwnq\" (UniqueName: \"kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq\") pod \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\" (UID: \"4c5eadde-8ce4-4009-9128-66d4d9657dfc\") " Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.065476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4c5eadde-8ce4-4009-9128-66d4d9657dfc" (UID: "4c5eadde-8ce4-4009-9128-66d4d9657dfc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.065927 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq" (OuterVolumeSpecName: "kube-api-access-tdwnq") pod "4c5eadde-8ce4-4009-9128-66d4d9657dfc" (UID: "4c5eadde-8ce4-4009-9128-66d4d9657dfc"). InnerVolumeSpecName "kube-api-access-tdwnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.092478 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory" (OuterVolumeSpecName: "inventory") pod "4c5eadde-8ce4-4009-9128-66d4d9657dfc" (UID: "4c5eadde-8ce4-4009-9128-66d4d9657dfc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.093942 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c5eadde-8ce4-4009-9128-66d4d9657dfc" (UID: "4c5eadde-8ce4-4009-9128-66d4d9657dfc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.116285 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "4c5eadde-8ce4-4009-9128-66d4d9657dfc" (UID: "4c5eadde-8ce4-4009-9128-66d4d9657dfc"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.160013 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.160471 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.160556 4960 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.160630 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4c5eadde-8ce4-4009-9128-66d4d9657dfc-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.160698 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdwnq\" (UniqueName: \"kubernetes.io/projected/4c5eadde-8ce4-4009-9128-66d4d9657dfc-kube-api-access-tdwnq\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.452631 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" event={"ID":"4c5eadde-8ce4-4009-9128-66d4d9657dfc","Type":"ContainerDied","Data":"b3a607404d60c7f33d6ede58738937ab8f824766f9a965c51d9b399d6901df58"} Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.452673 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3a607404d60c7f33d6ede58738937ab8f824766f9a965c51d9b399d6901df58" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.452760 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.540508 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5"] Dec 01 20:25:51 crc kubenswrapper[4960]: E1201 20:25:51.540919 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5eadde-8ce4-4009-9128-66d4d9657dfc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.540938 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5eadde-8ce4-4009-9128-66d4d9657dfc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.541195 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5eadde-8ce4-4009-9128-66d4d9657dfc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.542169 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.544793 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.545025 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.545296 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.545451 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.546936 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.547100 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.547268 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.559905 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5"] Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678166 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678285 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678354 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678420 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678481 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678518 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkwb5\" (UniqueName: \"kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.678579 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780434 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780529 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780579 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkwb5\" (UniqueName: \"kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780605 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780639 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780699 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780734 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780770 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.780826 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.781939 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.784885 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.784906 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.785347 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.785503 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.786386 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.786502 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.786525 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.797058 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkwb5\" (UniqueName: \"kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-stff5\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:51 crc kubenswrapper[4960]: I1201 20:25:51.859996 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:25:52 crc kubenswrapper[4960]: I1201 20:25:52.392911 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5"] Dec 01 20:25:52 crc kubenswrapper[4960]: I1201 20:25:52.462985 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" event={"ID":"e37fd55a-a5e3-4844-93b3-a22664d77bcd","Type":"ContainerStarted","Data":"b54fc042d013530d7d64e5b322824558c6c52b85bc9c25bcfcf450157e9e5cb3"} Dec 01 20:25:52 crc kubenswrapper[4960]: I1201 20:25:52.799779 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:52 crc kubenswrapper[4960]: I1201 20:25:52.849773 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:52 crc kubenswrapper[4960]: I1201 20:25:52.982907 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:53 crc kubenswrapper[4960]: I1201 20:25:53.028191 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:53 crc kubenswrapper[4960]: I1201 20:25:53.962918 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:54 crc kubenswrapper[4960]: I1201 20:25:54.498903 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" event={"ID":"e37fd55a-a5e3-4844-93b3-a22664d77bcd","Type":"ContainerStarted","Data":"003a23a59ac98c6d77dc0af9727dc23e585fe70d84de51544e822729ea6406ea"} Dec 01 20:25:54 crc kubenswrapper[4960]: I1201 20:25:54.499348 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8ncft" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" containerID="cri-o://70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d" gracePeriod=2 Dec 01 20:25:54 crc kubenswrapper[4960]: I1201 20:25:54.523703 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" podStartSLOduration=2.265748478 podStartE2EDuration="3.523686519s" podCreationTimestamp="2025-12-01 20:25:51 +0000 UTC" firstStartedPulling="2025-12-01 20:25:52.394715375 +0000 UTC m=+2787.682207044" lastFinishedPulling="2025-12-01 20:25:53.652653416 +0000 UTC m=+2788.940145085" observedRunningTime="2025-12-01 20:25:54.51954835 +0000 UTC m=+2789.807040019" watchObservedRunningTime="2025-12-01 20:25:54.523686519 +0000 UTC m=+2789.811178188" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.030980 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.191999 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkckb\" (UniqueName: \"kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb\") pod \"2711f399-db4b-435d-b9a7-61a50dcfa16e\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.192066 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content\") pod \"2711f399-db4b-435d-b9a7-61a50dcfa16e\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.192257 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities\") pod \"2711f399-db4b-435d-b9a7-61a50dcfa16e\" (UID: \"2711f399-db4b-435d-b9a7-61a50dcfa16e\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.193101 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities" (OuterVolumeSpecName: "utilities") pod "2711f399-db4b-435d-b9a7-61a50dcfa16e" (UID: "2711f399-db4b-435d-b9a7-61a50dcfa16e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.200450 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb" (OuterVolumeSpecName: "kube-api-access-rkckb") pod "2711f399-db4b-435d-b9a7-61a50dcfa16e" (UID: "2711f399-db4b-435d-b9a7-61a50dcfa16e"). InnerVolumeSpecName "kube-api-access-rkckb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.294755 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.294794 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkckb\" (UniqueName: \"kubernetes.io/projected/2711f399-db4b-435d-b9a7-61a50dcfa16e-kube-api-access-rkckb\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.326503 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2711f399-db4b-435d-b9a7-61a50dcfa16e" (UID: "2711f399-db4b-435d-b9a7-61a50dcfa16e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.365244 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.366968 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-74rqb" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="registry-server" containerID="cri-o://d3772d2f43f340e8ea4471c73c5c4f101479b92fdecc61a9abaef0f3ce9035c1" gracePeriod=2 Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.397268 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2711f399-db4b-435d-b9a7-61a50dcfa16e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.518406 4960 generic.go:334] "Generic (PLEG): container finished" podID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerID="70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d" exitCode=0 Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.518469 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerDied","Data":"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d"} Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.519316 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ncft" event={"ID":"2711f399-db4b-435d-b9a7-61a50dcfa16e","Type":"ContainerDied","Data":"f3eafc6df64776146af160735cab2dde6ca9a6752078f014e1b18576ff92efc2"} Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.518564 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ncft" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.520237 4960 scope.go:117] "RemoveContainer" containerID="70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.525893 4960 generic.go:334] "Generic (PLEG): container finished" podID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerID="d3772d2f43f340e8ea4471c73c5c4f101479b92fdecc61a9abaef0f3ce9035c1" exitCode=0 Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.526170 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerDied","Data":"d3772d2f43f340e8ea4471c73c5c4f101479b92fdecc61a9abaef0f3ce9035c1"} Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.591270 4960 scope.go:117] "RemoveContainer" containerID="435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.594271 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.617092 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8ncft"] Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.632800 4960 scope.go:117] "RemoveContainer" containerID="3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.680318 4960 scope.go:117] "RemoveContainer" containerID="70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d" Dec 01 20:25:55 crc kubenswrapper[4960]: E1201 20:25:55.684142 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d\": container with ID starting with 70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d not found: ID does not exist" containerID="70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.684182 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d"} err="failed to get container status \"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d\": rpc error: code = NotFound desc = could not find container \"70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d\": container with ID starting with 70a3b86e31a46c7cdef94281e7f7b11ff2718a536c0f785c672df1a22cde6c4d not found: ID does not exist" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.684233 4960 scope.go:117] "RemoveContainer" containerID="435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f" Dec 01 20:25:55 crc kubenswrapper[4960]: E1201 20:25:55.684591 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f\": container with ID starting with 435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f not found: ID does not exist" containerID="435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.684636 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f"} err="failed to get container status \"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f\": rpc error: code = NotFound desc = could not find container \"435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f\": container with ID starting with 435b0144e0d7b57fcd77e4c21a21142dd69be4ddf8290c1bc7b1d43e39cee77f not found: ID does not exist" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.684663 4960 scope.go:117] "RemoveContainer" containerID="3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514" Dec 01 20:25:55 crc kubenswrapper[4960]: E1201 20:25:55.684889 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514\": container with ID starting with 3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514 not found: ID does not exist" containerID="3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.684910 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514"} err="failed to get container status \"3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514\": rpc error: code = NotFound desc = could not find container \"3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514\": container with ID starting with 3824607cdf8c07cbd9ad927a1783d03d1c82cd93455bb5dce63f0ac832c1a514 not found: ID does not exist" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.747197 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.908036 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content\") pod \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.908286 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd65w\" (UniqueName: \"kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w\") pod \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.908344 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities\") pod \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\" (UID: \"eee7533d-39ce-42ac-bf5a-8ba62b121e09\") " Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.910073 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities" (OuterVolumeSpecName: "utilities") pod "eee7533d-39ce-42ac-bf5a-8ba62b121e09" (UID: "eee7533d-39ce-42ac-bf5a-8ba62b121e09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.913542 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w" (OuterVolumeSpecName: "kube-api-access-dd65w") pod "eee7533d-39ce-42ac-bf5a-8ba62b121e09" (UID: "eee7533d-39ce-42ac-bf5a-8ba62b121e09"). InnerVolumeSpecName "kube-api-access-dd65w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:25:55 crc kubenswrapper[4960]: I1201 20:25:55.953713 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eee7533d-39ce-42ac-bf5a-8ba62b121e09" (UID: "eee7533d-39ce-42ac-bf5a-8ba62b121e09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.011043 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd65w\" (UniqueName: \"kubernetes.io/projected/eee7533d-39ce-42ac-bf5a-8ba62b121e09-kube-api-access-dd65w\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.011083 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.011098 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee7533d-39ce-42ac-bf5a-8ba62b121e09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.539474 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-74rqb" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.539480 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-74rqb" event={"ID":"eee7533d-39ce-42ac-bf5a-8ba62b121e09","Type":"ContainerDied","Data":"e156d3e9af0f33f72e15ebb6078314c466da2d962209abdac37b3b5ab56aa7b5"} Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.541489 4960 scope.go:117] "RemoveContainer" containerID="d3772d2f43f340e8ea4471c73c5c4f101479b92fdecc61a9abaef0f3ce9035c1" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.576240 4960 scope.go:117] "RemoveContainer" containerID="0b6382163d0f89c3fca50c66c764e091c8e91f7cc2f3d8e1eb8d7bf5c78022c8" Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.581095 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.589332 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-74rqb"] Dec 01 20:25:56 crc kubenswrapper[4960]: I1201 20:25:56.605039 4960 scope.go:117] "RemoveContainer" containerID="fad8d58c5b7385db018cde92e5cadac22a356decdb69ccaef48f1d5147e75713" Dec 01 20:25:57 crc kubenswrapper[4960]: I1201 20:25:57.336339 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" path="/var/lib/kubelet/pods/2711f399-db4b-435d-b9a7-61a50dcfa16e/volumes" Dec 01 20:25:57 crc kubenswrapper[4960]: I1201 20:25:57.337097 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" path="/var/lib/kubelet/pods/eee7533d-39ce-42ac-bf5a-8ba62b121e09/volumes" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.854472 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.855893 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.855917 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.855988 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="extract-content" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856000 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="extract-content" Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.856019 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856031 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.856061 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="extract-utilities" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856073 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="extract-utilities" Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.856095 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="extract-content" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856107 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="extract-content" Dec 01 20:26:29 crc kubenswrapper[4960]: E1201 20:26:29.856163 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="extract-utilities" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856175 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="extract-utilities" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856550 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2711f399-db4b-435d-b9a7-61a50dcfa16e" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.856594 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee7533d-39ce-42ac-bf5a-8ba62b121e09" containerName="registry-server" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.866369 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.866884 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.978778 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.979455 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:29 crc kubenswrapper[4960]: I1201 20:26:29.979588 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl8t5\" (UniqueName: \"kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.081645 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.081700 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.081744 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl8t5\" (UniqueName: \"kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.082194 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.082231 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.119024 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl8t5\" (UniqueName: \"kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5\") pod \"community-operators-r8rfm\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.193691 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.737778 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:30 crc kubenswrapper[4960]: I1201 20:26:30.910688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerStarted","Data":"4dbd1109abb769cd599ff551cf84a55334a44eea776423fe509dbee3d0a56837"} Dec 01 20:26:31 crc kubenswrapper[4960]: I1201 20:26:31.925061 4960 generic.go:334] "Generic (PLEG): container finished" podID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerID="c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2" exitCode=0 Dec 01 20:26:31 crc kubenswrapper[4960]: I1201 20:26:31.925154 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerDied","Data":"c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2"} Dec 01 20:26:32 crc kubenswrapper[4960]: I1201 20:26:32.934984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerStarted","Data":"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3"} Dec 01 20:26:33 crc kubenswrapper[4960]: I1201 20:26:33.948831 4960 generic.go:334] "Generic (PLEG): container finished" podID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerID="bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3" exitCode=0 Dec 01 20:26:33 crc kubenswrapper[4960]: I1201 20:26:33.949196 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerDied","Data":"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3"} Dec 01 20:26:34 crc kubenswrapper[4960]: I1201 20:26:34.963100 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerStarted","Data":"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521"} Dec 01 20:26:34 crc kubenswrapper[4960]: I1201 20:26:34.979414 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r8rfm" podStartSLOduration=3.490679719 podStartE2EDuration="5.979396735s" podCreationTimestamp="2025-12-01 20:26:29 +0000 UTC" firstStartedPulling="2025-12-01 20:26:31.927932236 +0000 UTC m=+2827.215423945" lastFinishedPulling="2025-12-01 20:26:34.416649282 +0000 UTC m=+2829.704140961" observedRunningTime="2025-12-01 20:26:34.977181206 +0000 UTC m=+2830.264672885" watchObservedRunningTime="2025-12-01 20:26:34.979396735 +0000 UTC m=+2830.266888404" Dec 01 20:26:40 crc kubenswrapper[4960]: I1201 20:26:40.193932 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:40 crc kubenswrapper[4960]: I1201 20:26:40.194161 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:40 crc kubenswrapper[4960]: I1201 20:26:40.256253 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:41 crc kubenswrapper[4960]: I1201 20:26:41.068410 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:41 crc kubenswrapper[4960]: I1201 20:26:41.127689 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.032459 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r8rfm" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="registry-server" containerID="cri-o://f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521" gracePeriod=2 Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.532107 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.672873 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl8t5\" (UniqueName: \"kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5\") pod \"db15a2b2-74ac-4c81-896d-8be992660d0c\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.673035 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content\") pod \"db15a2b2-74ac-4c81-896d-8be992660d0c\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.673162 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities\") pod \"db15a2b2-74ac-4c81-896d-8be992660d0c\" (UID: \"db15a2b2-74ac-4c81-896d-8be992660d0c\") " Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.674783 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities" (OuterVolumeSpecName: "utilities") pod "db15a2b2-74ac-4c81-896d-8be992660d0c" (UID: "db15a2b2-74ac-4c81-896d-8be992660d0c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.682746 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5" (OuterVolumeSpecName: "kube-api-access-gl8t5") pod "db15a2b2-74ac-4c81-896d-8be992660d0c" (UID: "db15a2b2-74ac-4c81-896d-8be992660d0c"). InnerVolumeSpecName "kube-api-access-gl8t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.748698 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db15a2b2-74ac-4c81-896d-8be992660d0c" (UID: "db15a2b2-74ac-4c81-896d-8be992660d0c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.776572 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.776610 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl8t5\" (UniqueName: \"kubernetes.io/projected/db15a2b2-74ac-4c81-896d-8be992660d0c-kube-api-access-gl8t5\") on node \"crc\" DevicePath \"\"" Dec 01 20:26:43 crc kubenswrapper[4960]: I1201 20:26:43.776623 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db15a2b2-74ac-4c81-896d-8be992660d0c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.044783 4960 generic.go:334] "Generic (PLEG): container finished" podID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerID="f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521" exitCode=0 Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.044841 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r8rfm" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.044846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerDied","Data":"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521"} Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.044888 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r8rfm" event={"ID":"db15a2b2-74ac-4c81-896d-8be992660d0c","Type":"ContainerDied","Data":"4dbd1109abb769cd599ff551cf84a55334a44eea776423fe509dbee3d0a56837"} Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.044912 4960 scope.go:117] "RemoveContainer" containerID="f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.083189 4960 scope.go:117] "RemoveContainer" containerID="bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.090008 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.110332 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r8rfm"] Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.116016 4960 scope.go:117] "RemoveContainer" containerID="c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.162083 4960 scope.go:117] "RemoveContainer" containerID="f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521" Dec 01 20:26:44 crc kubenswrapper[4960]: E1201 20:26:44.162690 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521\": container with ID starting with f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521 not found: ID does not exist" containerID="f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.162741 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521"} err="failed to get container status \"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521\": rpc error: code = NotFound desc = could not find container \"f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521\": container with ID starting with f152ee4bf34020cdde291fe7c42d95b7241d1eca7b1f373a841f4432c7548521 not found: ID does not exist" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.162774 4960 scope.go:117] "RemoveContainer" containerID="bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3" Dec 01 20:26:44 crc kubenswrapper[4960]: E1201 20:26:44.163180 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3\": container with ID starting with bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3 not found: ID does not exist" containerID="bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.163212 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3"} err="failed to get container status \"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3\": rpc error: code = NotFound desc = could not find container \"bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3\": container with ID starting with bd4b4c017fe3497672a369dddd091047eea4a34ad9a19a0f003ee5e3070841c3 not found: ID does not exist" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.163234 4960 scope.go:117] "RemoveContainer" containerID="c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2" Dec 01 20:26:44 crc kubenswrapper[4960]: E1201 20:26:44.163461 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2\": container with ID starting with c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2 not found: ID does not exist" containerID="c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2" Dec 01 20:26:44 crc kubenswrapper[4960]: I1201 20:26:44.163493 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2"} err="failed to get container status \"c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2\": rpc error: code = NotFound desc = could not find container \"c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2\": container with ID starting with c5b8b505eab40934450c637e66769091245c5ade0db2fc4d92bea9c76d2c31f2 not found: ID does not exist" Dec 01 20:26:45 crc kubenswrapper[4960]: I1201 20:26:45.337827 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" path="/var/lib/kubelet/pods/db15a2b2-74ac-4c81-896d-8be992660d0c/volumes" Dec 01 20:27:13 crc kubenswrapper[4960]: I1201 20:27:13.870594 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:27:13 crc kubenswrapper[4960]: I1201 20:27:13.871259 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:27:43 crc kubenswrapper[4960]: I1201 20:27:43.870588 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:27:43 crc kubenswrapper[4960]: I1201 20:27:43.871179 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:28:13 crc kubenswrapper[4960]: I1201 20:28:13.871358 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:28:13 crc kubenswrapper[4960]: I1201 20:28:13.872035 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:28:13 crc kubenswrapper[4960]: I1201 20:28:13.872087 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:28:13 crc kubenswrapper[4960]: I1201 20:28:13.873065 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:28:13 crc kubenswrapper[4960]: I1201 20:28:13.873151 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5" gracePeriod=600 Dec 01 20:28:14 crc kubenswrapper[4960]: I1201 20:28:14.160730 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5" exitCode=0 Dec 01 20:28:14 crc kubenswrapper[4960]: I1201 20:28:14.160796 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5"} Dec 01 20:28:14 crc kubenswrapper[4960]: I1201 20:28:14.160841 4960 scope.go:117] "RemoveContainer" containerID="efaa180009dc96ca099705afcaa6bd957c693877f451ba2831abecf8431f7933" Dec 01 20:28:15 crc kubenswrapper[4960]: I1201 20:28:15.174223 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d"} Dec 01 20:29:01 crc kubenswrapper[4960]: I1201 20:29:01.731558 4960 generic.go:334] "Generic (PLEG): container finished" podID="e37fd55a-a5e3-4844-93b3-a22664d77bcd" containerID="003a23a59ac98c6d77dc0af9727dc23e585fe70d84de51544e822729ea6406ea" exitCode=0 Dec 01 20:29:01 crc kubenswrapper[4960]: I1201 20:29:01.731617 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" event={"ID":"e37fd55a-a5e3-4844-93b3-a22664d77bcd","Type":"ContainerDied","Data":"003a23a59ac98c6d77dc0af9727dc23e585fe70d84de51544e822729ea6406ea"} Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.339887 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373255 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373415 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373526 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373564 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373660 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373702 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373729 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkwb5\" (UniqueName: \"kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373778 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.373815 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1\") pod \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\" (UID: \"e37fd55a-a5e3-4844-93b3-a22664d77bcd\") " Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.381033 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5" (OuterVolumeSpecName: "kube-api-access-hkwb5") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "kube-api-access-hkwb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.382416 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.413169 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.420128 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.454386 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.454455 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.459724 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.461476 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.464359 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory" (OuterVolumeSpecName: "inventory") pod "e37fd55a-a5e3-4844-93b3-a22664d77bcd" (UID: "e37fd55a-a5e3-4844-93b3-a22664d77bcd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476335 4960 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476751 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476768 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkwb5\" (UniqueName: \"kubernetes.io/projected/e37fd55a-a5e3-4844-93b3-a22664d77bcd-kube-api-access-hkwb5\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476778 4960 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476789 4960 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476797 4960 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476806 4960 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476815 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37fd55a-a5e3-4844-93b3-a22664d77bcd-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.476827 4960 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e37fd55a-a5e3-4844-93b3-a22664d77bcd-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.753494 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" event={"ID":"e37fd55a-a5e3-4844-93b3-a22664d77bcd","Type":"ContainerDied","Data":"b54fc042d013530d7d64e5b322824558c6c52b85bc9c25bcfcf450157e9e5cb3"} Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.753550 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b54fc042d013530d7d64e5b322824558c6c52b85bc9c25bcfcf450157e9e5cb3" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.753584 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-stff5" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.951786 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc"] Dec 01 20:29:03 crc kubenswrapper[4960]: E1201 20:29:03.952336 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37fd55a-a5e3-4844-93b3-a22664d77bcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952357 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37fd55a-a5e3-4844-93b3-a22664d77bcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 20:29:03 crc kubenswrapper[4960]: E1201 20:29:03.952386 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="extract-utilities" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952396 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="extract-utilities" Dec 01 20:29:03 crc kubenswrapper[4960]: E1201 20:29:03.952408 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="registry-server" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952416 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="registry-server" Dec 01 20:29:03 crc kubenswrapper[4960]: E1201 20:29:03.952434 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="extract-content" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952442 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="extract-content" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952681 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="db15a2b2-74ac-4c81-896d-8be992660d0c" containerName="registry-server" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.952699 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37fd55a-a5e3-4844-93b3-a22664d77bcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.953528 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.958577 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.959868 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.959987 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mxdld" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.960032 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.959996 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.965050 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc"] Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.987944 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988060 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988104 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbkr\" (UniqueName: \"kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988269 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988418 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988675 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:03 crc kubenswrapper[4960]: I1201 20:29:03.988819 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090371 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090434 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090458 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbkr\" (UniqueName: \"kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090502 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090538 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090612 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.090653 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.094161 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.094173 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.095136 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.096514 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.096719 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.096979 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.107738 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbkr\" (UniqueName: \"kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.284396 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:29:04 crc kubenswrapper[4960]: I1201 20:29:04.931476 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc"] Dec 01 20:29:05 crc kubenswrapper[4960]: I1201 20:29:05.782740 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" event={"ID":"00d8a61e-f3a9-4f45-9ff3-a196f814de12","Type":"ContainerStarted","Data":"c1ff715e535d00c062545c0e1b85dcc7beeb6ae95eba80e2615b4d6807bfc7f0"} Dec 01 20:29:06 crc kubenswrapper[4960]: I1201 20:29:06.799201 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" event={"ID":"00d8a61e-f3a9-4f45-9ff3-a196f814de12","Type":"ContainerStarted","Data":"c8a0d7726ce5ad328ca13b8039c0c9d79f74367c815731fc7f8a6ddb98f6ee72"} Dec 01 20:29:06 crc kubenswrapper[4960]: I1201 20:29:06.837381 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" podStartSLOduration=3.185548585 podStartE2EDuration="3.837350678s" podCreationTimestamp="2025-12-01 20:29:03 +0000 UTC" firstStartedPulling="2025-12-01 20:29:04.928711388 +0000 UTC m=+2980.216203067" lastFinishedPulling="2025-12-01 20:29:05.580513481 +0000 UTC m=+2980.868005160" observedRunningTime="2025-12-01 20:29:06.821151294 +0000 UTC m=+2982.108642993" watchObservedRunningTime="2025-12-01 20:29:06.837350678 +0000 UTC m=+2982.124842387" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.174724 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd"] Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.177976 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.182786 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.183075 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.227258 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd"] Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.296551 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.296642 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.296706 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thwv5\" (UniqueName: \"kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.398275 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.398346 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.398379 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thwv5\" (UniqueName: \"kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.399998 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.406777 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.419740 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thwv5\" (UniqueName: \"kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5\") pod \"collect-profiles-29410350-p7cqd\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:00 crc kubenswrapper[4960]: I1201 20:30:00.525361 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:01 crc kubenswrapper[4960]: I1201 20:30:01.024866 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd"] Dec 01 20:30:01 crc kubenswrapper[4960]: I1201 20:30:01.684494 4960 generic.go:334] "Generic (PLEG): container finished" podID="4b58a9bc-2600-4b34-a3bf-81c899a03ac7" containerID="4c2c76e92b167693981a26661a46d0a95434b19eda9e172f33129cfdd26a5df7" exitCode=0 Dec 01 20:30:01 crc kubenswrapper[4960]: I1201 20:30:01.684710 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" event={"ID":"4b58a9bc-2600-4b34-a3bf-81c899a03ac7","Type":"ContainerDied","Data":"4c2c76e92b167693981a26661a46d0a95434b19eda9e172f33129cfdd26a5df7"} Dec 01 20:30:01 crc kubenswrapper[4960]: I1201 20:30:01.684779 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" event={"ID":"4b58a9bc-2600-4b34-a3bf-81c899a03ac7","Type":"ContainerStarted","Data":"e0f5e68e7af442140ec0497a4aa3891f573db77ce5a374832e520175508bbd66"} Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.103865 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.258973 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume\") pod \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.259188 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume\") pod \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.259298 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thwv5\" (UniqueName: \"kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5\") pod \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\" (UID: \"4b58a9bc-2600-4b34-a3bf-81c899a03ac7\") " Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.260014 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b58a9bc-2600-4b34-a3bf-81c899a03ac7" (UID: "4b58a9bc-2600-4b34-a3bf-81c899a03ac7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.261222 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.264857 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5" (OuterVolumeSpecName: "kube-api-access-thwv5") pod "4b58a9bc-2600-4b34-a3bf-81c899a03ac7" (UID: "4b58a9bc-2600-4b34-a3bf-81c899a03ac7"). InnerVolumeSpecName "kube-api-access-thwv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.266386 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b58a9bc-2600-4b34-a3bf-81c899a03ac7" (UID: "4b58a9bc-2600-4b34-a3bf-81c899a03ac7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.363822 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.363868 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thwv5\" (UniqueName: \"kubernetes.io/projected/4b58a9bc-2600-4b34-a3bf-81c899a03ac7-kube-api-access-thwv5\") on node \"crc\" DevicePath \"\"" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.703457 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" event={"ID":"4b58a9bc-2600-4b34-a3bf-81c899a03ac7","Type":"ContainerDied","Data":"e0f5e68e7af442140ec0497a4aa3891f573db77ce5a374832e520175508bbd66"} Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.703859 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f5e68e7af442140ec0497a4aa3891f573db77ce5a374832e520175508bbd66" Dec 01 20:30:03 crc kubenswrapper[4960]: I1201 20:30:03.703627 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410350-p7cqd" Dec 01 20:30:04 crc kubenswrapper[4960]: I1201 20:30:04.190619 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8"] Dec 01 20:30:04 crc kubenswrapper[4960]: I1201 20:30:04.198705 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410305-q7jp8"] Dec 01 20:30:05 crc kubenswrapper[4960]: I1201 20:30:05.345897 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1beb3eb3-cb74-437d-9c99-faf228d0e172" path="/var/lib/kubelet/pods/1beb3eb3-cb74-437d-9c99-faf228d0e172/volumes" Dec 01 20:30:16 crc kubenswrapper[4960]: I1201 20:30:16.068162 4960 scope.go:117] "RemoveContainer" containerID="e87ae19577a2513888ea613216aa446c11b0d66c07788f7d54fec47832eb1dc6" Dec 01 20:30:43 crc kubenswrapper[4960]: I1201 20:30:43.870984 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:30:43 crc kubenswrapper[4960]: I1201 20:30:43.871554 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:31:13 crc kubenswrapper[4960]: I1201 20:31:13.883819 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:31:13 crc kubenswrapper[4960]: I1201 20:31:13.884346 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:31:43 crc kubenswrapper[4960]: I1201 20:31:43.871365 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:31:43 crc kubenswrapper[4960]: I1201 20:31:43.871908 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:31:43 crc kubenswrapper[4960]: I1201 20:31:43.871960 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:31:43 crc kubenswrapper[4960]: I1201 20:31:43.872709 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:31:43 crc kubenswrapper[4960]: I1201 20:31:43.872758 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" gracePeriod=600 Dec 01 20:31:44 crc kubenswrapper[4960]: E1201 20:31:44.002981 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:31:44 crc kubenswrapper[4960]: I1201 20:31:44.826374 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" exitCode=0 Dec 01 20:31:44 crc kubenswrapper[4960]: I1201 20:31:44.826511 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d"} Dec 01 20:31:44 crc kubenswrapper[4960]: I1201 20:31:44.826779 4960 scope.go:117] "RemoveContainer" containerID="e5a264d54f5353b0337ad79df143c730b3c8d35cfe4f40f4d1ab200aabf16ac5" Dec 01 20:31:44 crc kubenswrapper[4960]: I1201 20:31:44.827573 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:31:44 crc kubenswrapper[4960]: E1201 20:31:44.827894 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:31:48 crc kubenswrapper[4960]: I1201 20:31:48.867439 4960 generic.go:334] "Generic (PLEG): container finished" podID="00d8a61e-f3a9-4f45-9ff3-a196f814de12" containerID="c8a0d7726ce5ad328ca13b8039c0c9d79f74367c815731fc7f8a6ddb98f6ee72" exitCode=0 Dec 01 20:31:48 crc kubenswrapper[4960]: I1201 20:31:48.867984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" event={"ID":"00d8a61e-f3a9-4f45-9ff3-a196f814de12","Type":"ContainerDied","Data":"c8a0d7726ce5ad328ca13b8039c0c9d79f74367c815731fc7f8a6ddb98f6ee72"} Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.414976 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.476603 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477501 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477585 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477760 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kbkr\" (UniqueName: \"kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477805 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477874 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.477896 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1\") pod \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\" (UID: \"00d8a61e-f3a9-4f45-9ff3-a196f814de12\") " Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.482983 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.484977 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr" (OuterVolumeSpecName: "kube-api-access-4kbkr") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "kube-api-access-4kbkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.512523 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory" (OuterVolumeSpecName: "inventory") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.513250 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.516440 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.518168 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.532208 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00d8a61e-f3a9-4f45-9ff3-a196f814de12" (UID: "00d8a61e-f3a9-4f45-9ff3-a196f814de12"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579769 4960 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579803 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579813 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579825 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kbkr\" (UniqueName: \"kubernetes.io/projected/00d8a61e-f3a9-4f45-9ff3-a196f814de12-kube-api-access-4kbkr\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579836 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579845 4960 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-inventory\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.579854 4960 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/00d8a61e-f3a9-4f45-9ff3-a196f814de12-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.890246 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" event={"ID":"00d8a61e-f3a9-4f45-9ff3-a196f814de12","Type":"ContainerDied","Data":"c1ff715e535d00c062545c0e1b85dcc7beeb6ae95eba80e2615b4d6807bfc7f0"} Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.890309 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1ff715e535d00c062545c0e1b85dcc7beeb6ae95eba80e2615b4d6807bfc7f0" Dec 01 20:31:50 crc kubenswrapper[4960]: I1201 20:31:50.890335 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc" Dec 01 20:31:57 crc kubenswrapper[4960]: I1201 20:31:57.324814 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:31:57 crc kubenswrapper[4960]: E1201 20:31:57.325532 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:32:09 crc kubenswrapper[4960]: I1201 20:32:09.324244 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:32:09 crc kubenswrapper[4960]: E1201 20:32:09.324994 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:32:22 crc kubenswrapper[4960]: I1201 20:32:22.324690 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:32:22 crc kubenswrapper[4960]: E1201 20:32:22.325555 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:32:35 crc kubenswrapper[4960]: I1201 20:32:35.332262 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:32:35 crc kubenswrapper[4960]: E1201 20:32:35.333252 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:32:49 crc kubenswrapper[4960]: I1201 20:32:49.324763 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:32:49 crc kubenswrapper[4960]: E1201 20:32:49.325715 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:33:04 crc kubenswrapper[4960]: I1201 20:33:04.324058 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:33:04 crc kubenswrapper[4960]: E1201 20:33:04.325829 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:33:15 crc kubenswrapper[4960]: I1201 20:33:15.330782 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:33:15 crc kubenswrapper[4960]: E1201 20:33:15.331617 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.546258 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 20:33:23 crc kubenswrapper[4960]: E1201 20:33:23.547837 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b58a9bc-2600-4b34-a3bf-81c899a03ac7" containerName="collect-profiles" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.547857 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b58a9bc-2600-4b34-a3bf-81c899a03ac7" containerName="collect-profiles" Dec 01 20:33:23 crc kubenswrapper[4960]: E1201 20:33:23.547879 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d8a61e-f3a9-4f45-9ff3-a196f814de12" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.547888 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d8a61e-f3a9-4f45-9ff3-a196f814de12" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.548180 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d8a61e-f3a9-4f45-9ff3-a196f814de12" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.548221 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b58a9bc-2600-4b34-a3bf-81c899a03ac7" containerName="collect-profiles" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.549187 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.550769 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-r2cdg" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.551343 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.551485 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.555409 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.557473 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.632795 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.632851 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.632914 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.632946 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.633190 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwns4\" (UniqueName: \"kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.633236 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.633446 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.633537 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.633645 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735625 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwns4\" (UniqueName: \"kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735681 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735761 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735812 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735862 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735930 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.735961 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.736019 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.736045 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.736730 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.737187 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.737805 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.737912 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.738291 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.744189 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.744381 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.745731 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.761614 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwns4\" (UniqueName: \"kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.767842 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " pod="openstack/tempest-tests-tempest" Dec 01 20:33:23 crc kubenswrapper[4960]: I1201 20:33:23.868802 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 20:33:24 crc kubenswrapper[4960]: I1201 20:33:24.366992 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 01 20:33:24 crc kubenswrapper[4960]: I1201 20:33:24.368810 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:33:24 crc kubenswrapper[4960]: I1201 20:33:24.524282 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d50fcead-df5c-40a0-b60f-4c195869572d","Type":"ContainerStarted","Data":"e2061bcfd548ad6bb478f0ff42b7df6ec6326e0944ed282d67015aa7e26e72e0"} Dec 01 20:33:27 crc kubenswrapper[4960]: I1201 20:33:27.326575 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:33:27 crc kubenswrapper[4960]: E1201 20:33:27.327506 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:33:42 crc kubenswrapper[4960]: I1201 20:33:42.324954 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:33:42 crc kubenswrapper[4960]: E1201 20:33:42.326068 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:33:52 crc kubenswrapper[4960]: E1201 20:33:52.212023 4960 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 01 20:33:52 crc kubenswrapper[4960]: E1201 20:33:52.213189 4960 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwns4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d50fcead-df5c-40a0-b60f-4c195869572d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 01 20:33:52 crc kubenswrapper[4960]: E1201 20:33:52.214481 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d50fcead-df5c-40a0-b60f-4c195869572d" Dec 01 20:33:52 crc kubenswrapper[4960]: E1201 20:33:52.846343 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d50fcead-df5c-40a0-b60f-4c195869572d" Dec 01 20:33:55 crc kubenswrapper[4960]: I1201 20:33:55.339831 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:33:55 crc kubenswrapper[4960]: E1201 20:33:55.340659 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:34:08 crc kubenswrapper[4960]: I1201 20:34:08.401631 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 01 20:34:09 crc kubenswrapper[4960]: I1201 20:34:09.324925 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:34:09 crc kubenswrapper[4960]: E1201 20:34:09.325801 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:34:10 crc kubenswrapper[4960]: I1201 20:34:10.090811 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d50fcead-df5c-40a0-b60f-4c195869572d","Type":"ContainerStarted","Data":"7177c7e65162a79f0fc380fe26e43880d1e43e8cf3cdd9214fc528d19099b7de"} Dec 01 20:34:10 crc kubenswrapper[4960]: I1201 20:34:10.121795 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.092548156 podStartE2EDuration="48.121773864s" podCreationTimestamp="2025-12-01 20:33:22 +0000 UTC" firstStartedPulling="2025-12-01 20:33:24.368590958 +0000 UTC m=+3239.656082627" lastFinishedPulling="2025-12-01 20:34:08.397816626 +0000 UTC m=+3283.685308335" observedRunningTime="2025-12-01 20:34:10.115454858 +0000 UTC m=+3285.402946567" watchObservedRunningTime="2025-12-01 20:34:10.121773864 +0000 UTC m=+3285.409265533" Dec 01 20:34:20 crc kubenswrapper[4960]: I1201 20:34:20.324999 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:34:20 crc kubenswrapper[4960]: E1201 20:34:20.326379 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:34:34 crc kubenswrapper[4960]: I1201 20:34:34.324663 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:34:34 crc kubenswrapper[4960]: E1201 20:34:34.326529 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:34:45 crc kubenswrapper[4960]: I1201 20:34:45.332822 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:34:45 crc kubenswrapper[4960]: E1201 20:34:45.333815 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:34:57 crc kubenswrapper[4960]: I1201 20:34:57.324364 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:34:57 crc kubenswrapper[4960]: E1201 20:34:57.325327 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:35:11 crc kubenswrapper[4960]: I1201 20:35:11.325026 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:35:11 crc kubenswrapper[4960]: E1201 20:35:11.325930 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.908522 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.911841 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.922084 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.995432 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.995507 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:24 crc kubenswrapper[4960]: I1201 20:35:24.995587 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7k8m\" (UniqueName: \"kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.098185 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.098275 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.098336 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7k8m\" (UniqueName: \"kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.098672 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.098752 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.126326 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7k8m\" (UniqueName: \"kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m\") pod \"redhat-marketplace-glr5d\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.235974 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.334642 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:35:25 crc kubenswrapper[4960]: E1201 20:35:25.335285 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.755837 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:25 crc kubenswrapper[4960]: I1201 20:35:25.927784 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerStarted","Data":"8401d1b2b02138be2bb6b9590b4496aaa78e630aaa053b256b02d1c7348b9122"} Dec 01 20:35:26 crc kubenswrapper[4960]: I1201 20:35:26.937903 4960 generic.go:334] "Generic (PLEG): container finished" podID="078fc35b-0326-4840-9a46-d6af376cddcc" containerID="62fd19b25344e79b22684b6072b734bdffcf0b5c766fb3821f0039e8da1dcf9f" exitCode=0 Dec 01 20:35:26 crc kubenswrapper[4960]: I1201 20:35:26.937964 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerDied","Data":"62fd19b25344e79b22684b6072b734bdffcf0b5c766fb3821f0039e8da1dcf9f"} Dec 01 20:35:27 crc kubenswrapper[4960]: I1201 20:35:27.973000 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerStarted","Data":"913b3f902337b5c8c40903ec94f631cb1b6d2c9f2a8b72a100e8c2db5af9c634"} Dec 01 20:35:28 crc kubenswrapper[4960]: I1201 20:35:28.985080 4960 generic.go:334] "Generic (PLEG): container finished" podID="078fc35b-0326-4840-9a46-d6af376cddcc" containerID="913b3f902337b5c8c40903ec94f631cb1b6d2c9f2a8b72a100e8c2db5af9c634" exitCode=0 Dec 01 20:35:28 crc kubenswrapper[4960]: I1201 20:35:28.985147 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerDied","Data":"913b3f902337b5c8c40903ec94f631cb1b6d2c9f2a8b72a100e8c2db5af9c634"} Dec 01 20:35:32 crc kubenswrapper[4960]: I1201 20:35:32.018177 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerStarted","Data":"78e71a19a645c9b5f784c6bf524bbfdff7f63664a1a10f6551966f31a0d066b1"} Dec 01 20:35:32 crc kubenswrapper[4960]: I1201 20:35:32.041824 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-glr5d" podStartSLOduration=3.941684901 podStartE2EDuration="8.041799481s" podCreationTimestamp="2025-12-01 20:35:24 +0000 UTC" firstStartedPulling="2025-12-01 20:35:26.941731804 +0000 UTC m=+3362.229223473" lastFinishedPulling="2025-12-01 20:35:31.041846374 +0000 UTC m=+3366.329338053" observedRunningTime="2025-12-01 20:35:32.034526024 +0000 UTC m=+3367.322017713" watchObservedRunningTime="2025-12-01 20:35:32.041799481 +0000 UTC m=+3367.329291150" Dec 01 20:35:35 crc kubenswrapper[4960]: I1201 20:35:35.236260 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:35 crc kubenswrapper[4960]: I1201 20:35:35.236919 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:35 crc kubenswrapper[4960]: I1201 20:35:35.287857 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:36 crc kubenswrapper[4960]: I1201 20:35:36.102252 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:36 crc kubenswrapper[4960]: I1201 20:35:36.154493 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:38 crc kubenswrapper[4960]: I1201 20:35:38.070473 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-glr5d" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="registry-server" containerID="cri-o://78e71a19a645c9b5f784c6bf524bbfdff7f63664a1a10f6551966f31a0d066b1" gracePeriod=2 Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.086286 4960 generic.go:334] "Generic (PLEG): container finished" podID="078fc35b-0326-4840-9a46-d6af376cddcc" containerID="78e71a19a645c9b5f784c6bf524bbfdff7f63664a1a10f6551966f31a0d066b1" exitCode=0 Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.086510 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerDied","Data":"78e71a19a645c9b5f784c6bf524bbfdff7f63664a1a10f6551966f31a0d066b1"} Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.335766 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.412068 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content\") pod \"078fc35b-0326-4840-9a46-d6af376cddcc\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.412223 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities\") pod \"078fc35b-0326-4840-9a46-d6af376cddcc\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.412283 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7k8m\" (UniqueName: \"kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m\") pod \"078fc35b-0326-4840-9a46-d6af376cddcc\" (UID: \"078fc35b-0326-4840-9a46-d6af376cddcc\") " Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.413104 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities" (OuterVolumeSpecName: "utilities") pod "078fc35b-0326-4840-9a46-d6af376cddcc" (UID: "078fc35b-0326-4840-9a46-d6af376cddcc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.418444 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m" (OuterVolumeSpecName: "kube-api-access-r7k8m") pod "078fc35b-0326-4840-9a46-d6af376cddcc" (UID: "078fc35b-0326-4840-9a46-d6af376cddcc"). InnerVolumeSpecName "kube-api-access-r7k8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.434191 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "078fc35b-0326-4840-9a46-d6af376cddcc" (UID: "078fc35b-0326-4840-9a46-d6af376cddcc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.514963 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7k8m\" (UniqueName: \"kubernetes.io/projected/078fc35b-0326-4840-9a46-d6af376cddcc-kube-api-access-r7k8m\") on node \"crc\" DevicePath \"\"" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.515312 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:35:39 crc kubenswrapper[4960]: I1201 20:35:39.515324 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/078fc35b-0326-4840-9a46-d6af376cddcc-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.098018 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-glr5d" event={"ID":"078fc35b-0326-4840-9a46-d6af376cddcc","Type":"ContainerDied","Data":"8401d1b2b02138be2bb6b9590b4496aaa78e630aaa053b256b02d1c7348b9122"} Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.098078 4960 scope.go:117] "RemoveContainer" containerID="78e71a19a645c9b5f784c6bf524bbfdff7f63664a1a10f6551966f31a0d066b1" Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.098084 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-glr5d" Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.125286 4960 scope.go:117] "RemoveContainer" containerID="913b3f902337b5c8c40903ec94f631cb1b6d2c9f2a8b72a100e8c2db5af9c634" Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.141124 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.151308 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-glr5d"] Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.155286 4960 scope.go:117] "RemoveContainer" containerID="62fd19b25344e79b22684b6072b734bdffcf0b5c766fb3821f0039e8da1dcf9f" Dec 01 20:35:40 crc kubenswrapper[4960]: I1201 20:35:40.326413 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:35:40 crc kubenswrapper[4960]: E1201 20:35:40.327690 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:35:41 crc kubenswrapper[4960]: I1201 20:35:41.336897 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" path="/var/lib/kubelet/pods/078fc35b-0326-4840-9a46-d6af376cddcc/volumes" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.826377 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:35:50 crc kubenswrapper[4960]: E1201 20:35:50.827674 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="extract-content" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.827698 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="extract-content" Dec 01 20:35:50 crc kubenswrapper[4960]: E1201 20:35:50.827771 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="extract-utilities" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.827786 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="extract-utilities" Dec 01 20:35:50 crc kubenswrapper[4960]: E1201 20:35:50.827807 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="registry-server" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.827819 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="registry-server" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.828165 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="078fc35b-0326-4840-9a46-d6af376cddcc" containerName="registry-server" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.830790 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.852269 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.955688 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.956066 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6x9s\" (UniqueName: \"kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:50 crc kubenswrapper[4960]: I1201 20:35:50.956785 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.059030 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.059103 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.059141 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6x9s\" (UniqueName: \"kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.059675 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.059698 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.086370 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6x9s\" (UniqueName: \"kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s\") pod \"certified-operators-kv4mq\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.174951 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:35:51 crc kubenswrapper[4960]: I1201 20:35:51.743501 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:35:52 crc kubenswrapper[4960]: I1201 20:35:52.240651 4960 generic.go:334] "Generic (PLEG): container finished" podID="54c3fa3b-d468-46ff-a449-9d777d487119" containerID="ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4" exitCode=0 Dec 01 20:35:52 crc kubenswrapper[4960]: I1201 20:35:52.240696 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerDied","Data":"ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4"} Dec 01 20:35:52 crc kubenswrapper[4960]: I1201 20:35:52.240720 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerStarted","Data":"567788140b38415e9f1b750ca01da39d41fa8170a7294875b89c57e2acfb17ff"} Dec 01 20:35:53 crc kubenswrapper[4960]: I1201 20:35:53.256804 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerStarted","Data":"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953"} Dec 01 20:35:54 crc kubenswrapper[4960]: I1201 20:35:54.272505 4960 generic.go:334] "Generic (PLEG): container finished" podID="54c3fa3b-d468-46ff-a449-9d777d487119" containerID="3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953" exitCode=0 Dec 01 20:35:54 crc kubenswrapper[4960]: I1201 20:35:54.272561 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerDied","Data":"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953"} Dec 01 20:35:55 crc kubenswrapper[4960]: I1201 20:35:55.333198 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:35:55 crc kubenswrapper[4960]: E1201 20:35:55.333898 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:35:56 crc kubenswrapper[4960]: I1201 20:35:56.300323 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerStarted","Data":"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28"} Dec 01 20:35:56 crc kubenswrapper[4960]: I1201 20:35:56.329007 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kv4mq" podStartSLOduration=3.363619477 podStartE2EDuration="6.328989446s" podCreationTimestamp="2025-12-01 20:35:50 +0000 UTC" firstStartedPulling="2025-12-01 20:35:52.242366135 +0000 UTC m=+3387.529857804" lastFinishedPulling="2025-12-01 20:35:55.207736104 +0000 UTC m=+3390.495227773" observedRunningTime="2025-12-01 20:35:56.319374577 +0000 UTC m=+3391.606866246" watchObservedRunningTime="2025-12-01 20:35:56.328989446 +0000 UTC m=+3391.616481115" Dec 01 20:36:01 crc kubenswrapper[4960]: I1201 20:36:01.176499 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:01 crc kubenswrapper[4960]: I1201 20:36:01.177089 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:01 crc kubenswrapper[4960]: I1201 20:36:01.238347 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:01 crc kubenswrapper[4960]: I1201 20:36:01.418101 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:01 crc kubenswrapper[4960]: I1201 20:36:01.483536 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:36:03 crc kubenswrapper[4960]: I1201 20:36:03.377399 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kv4mq" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="registry-server" containerID="cri-o://174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28" gracePeriod=2 Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.052631 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.165026 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content\") pod \"54c3fa3b-d468-46ff-a449-9d777d487119\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.165181 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6x9s\" (UniqueName: \"kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s\") pod \"54c3fa3b-d468-46ff-a449-9d777d487119\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.165216 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities\") pod \"54c3fa3b-d468-46ff-a449-9d777d487119\" (UID: \"54c3fa3b-d468-46ff-a449-9d777d487119\") " Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.165985 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities" (OuterVolumeSpecName: "utilities") pod "54c3fa3b-d468-46ff-a449-9d777d487119" (UID: "54c3fa3b-d468-46ff-a449-9d777d487119"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.166754 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.172366 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s" (OuterVolumeSpecName: "kube-api-access-t6x9s") pod "54c3fa3b-d468-46ff-a449-9d777d487119" (UID: "54c3fa3b-d468-46ff-a449-9d777d487119"). InnerVolumeSpecName "kube-api-access-t6x9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.229730 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54c3fa3b-d468-46ff-a449-9d777d487119" (UID: "54c3fa3b-d468-46ff-a449-9d777d487119"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.268277 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54c3fa3b-d468-46ff-a449-9d777d487119-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.268324 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6x9s\" (UniqueName: \"kubernetes.io/projected/54c3fa3b-d468-46ff-a449-9d777d487119-kube-api-access-t6x9s\") on node \"crc\" DevicePath \"\"" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.393539 4960 generic.go:334] "Generic (PLEG): container finished" podID="54c3fa3b-d468-46ff-a449-9d777d487119" containerID="174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28" exitCode=0 Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.393748 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerDied","Data":"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28"} Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.394149 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kv4mq" event={"ID":"54c3fa3b-d468-46ff-a449-9d777d487119","Type":"ContainerDied","Data":"567788140b38415e9f1b750ca01da39d41fa8170a7294875b89c57e2acfb17ff"} Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.394183 4960 scope.go:117] "RemoveContainer" containerID="174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.393869 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kv4mq" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.430407 4960 scope.go:117] "RemoveContainer" containerID="3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.466626 4960 scope.go:117] "RemoveContainer" containerID="ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.472528 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.486928 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kv4mq"] Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.527124 4960 scope.go:117] "RemoveContainer" containerID="174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28" Dec 01 20:36:04 crc kubenswrapper[4960]: E1201 20:36:04.527560 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28\": container with ID starting with 174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28 not found: ID does not exist" containerID="174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.527586 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28"} err="failed to get container status \"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28\": rpc error: code = NotFound desc = could not find container \"174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28\": container with ID starting with 174fcc294b941f82044fb0fede34b6533de67b81ba62ad137bfddfec46513e28 not found: ID does not exist" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.527606 4960 scope.go:117] "RemoveContainer" containerID="3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953" Dec 01 20:36:04 crc kubenswrapper[4960]: E1201 20:36:04.527892 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953\": container with ID starting with 3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953 not found: ID does not exist" containerID="3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.527915 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953"} err="failed to get container status \"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953\": rpc error: code = NotFound desc = could not find container \"3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953\": container with ID starting with 3de03c111e8d233a54d300299fb912e404d0ffaade8921f29a3e64645e6c9953 not found: ID does not exist" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.527929 4960 scope.go:117] "RemoveContainer" containerID="ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4" Dec 01 20:36:04 crc kubenswrapper[4960]: E1201 20:36:04.528183 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4\": container with ID starting with ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4 not found: ID does not exist" containerID="ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4" Dec 01 20:36:04 crc kubenswrapper[4960]: I1201 20:36:04.528205 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4"} err="failed to get container status \"ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4\": rpc error: code = NotFound desc = could not find container \"ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4\": container with ID starting with ad38495a0319485099aa7a60cf6804e1a0a38b00dc6cc514cad880b8ec55d3b4 not found: ID does not exist" Dec 01 20:36:05 crc kubenswrapper[4960]: I1201 20:36:05.339726 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" path="/var/lib/kubelet/pods/54c3fa3b-d468-46ff-a449-9d777d487119/volumes" Dec 01 20:36:08 crc kubenswrapper[4960]: I1201 20:36:08.324245 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:36:08 crc kubenswrapper[4960]: E1201 20:36:08.324846 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:36:20 crc kubenswrapper[4960]: I1201 20:36:20.324283 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:36:20 crc kubenswrapper[4960]: E1201 20:36:20.325206 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:36:33 crc kubenswrapper[4960]: I1201 20:36:33.324924 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:36:33 crc kubenswrapper[4960]: E1201 20:36:33.325790 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:36:47 crc kubenswrapper[4960]: I1201 20:36:47.324105 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:36:47 crc kubenswrapper[4960]: I1201 20:36:47.797434 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c"} Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.775400 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:36:50 crc kubenswrapper[4960]: E1201 20:36:50.776529 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="extract-content" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.776547 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="extract-content" Dec 01 20:36:50 crc kubenswrapper[4960]: E1201 20:36:50.776574 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="extract-utilities" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.776582 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="extract-utilities" Dec 01 20:36:50 crc kubenswrapper[4960]: E1201 20:36:50.776602 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="registry-server" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.776612 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="registry-server" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.776870 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="54c3fa3b-d468-46ff-a449-9d777d487119" containerName="registry-server" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.779483 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.788368 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.886763 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.886876 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.886955 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj9gc\" (UniqueName: \"kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.989009 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.989131 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj9gc\" (UniqueName: \"kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.989206 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.989612 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:50 crc kubenswrapper[4960]: I1201 20:36:50.989634 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:51 crc kubenswrapper[4960]: I1201 20:36:51.020005 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj9gc\" (UniqueName: \"kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc\") pod \"community-operators-h2c6d\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:51 crc kubenswrapper[4960]: I1201 20:36:51.108848 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:36:52 crc kubenswrapper[4960]: I1201 20:36:51.674652 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:36:52 crc kubenswrapper[4960]: W1201 20:36:51.686844 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33234aa0_f048_4e27_b8fb_9c351bb6b2cd.slice/crio-24b9278dbe5202479a7562eb8822f300d363d58c9eebb59a1380a87368e84db4 WatchSource:0}: Error finding container 24b9278dbe5202479a7562eb8822f300d363d58c9eebb59a1380a87368e84db4: Status 404 returned error can't find the container with id 24b9278dbe5202479a7562eb8822f300d363d58c9eebb59a1380a87368e84db4 Dec 01 20:36:52 crc kubenswrapper[4960]: I1201 20:36:51.851764 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerStarted","Data":"24b9278dbe5202479a7562eb8822f300d363d58c9eebb59a1380a87368e84db4"} Dec 01 20:36:52 crc kubenswrapper[4960]: I1201 20:36:52.867804 4960 generic.go:334] "Generic (PLEG): container finished" podID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerID="5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3" exitCode=0 Dec 01 20:36:52 crc kubenswrapper[4960]: I1201 20:36:52.868083 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerDied","Data":"5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3"} Dec 01 20:36:53 crc kubenswrapper[4960]: I1201 20:36:53.881217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerStarted","Data":"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c"} Dec 01 20:36:54 crc kubenswrapper[4960]: I1201 20:36:54.896308 4960 generic.go:334] "Generic (PLEG): container finished" podID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerID="57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c" exitCode=0 Dec 01 20:36:54 crc kubenswrapper[4960]: I1201 20:36:54.896477 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerDied","Data":"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c"} Dec 01 20:36:55 crc kubenswrapper[4960]: I1201 20:36:55.929633 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerStarted","Data":"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b"} Dec 01 20:36:55 crc kubenswrapper[4960]: I1201 20:36:55.954318 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h2c6d" podStartSLOduration=3.366571509 podStartE2EDuration="5.954298566s" podCreationTimestamp="2025-12-01 20:36:50 +0000 UTC" firstStartedPulling="2025-12-01 20:36:52.87151649 +0000 UTC m=+3448.159008159" lastFinishedPulling="2025-12-01 20:36:55.459243537 +0000 UTC m=+3450.746735216" observedRunningTime="2025-12-01 20:36:55.951254441 +0000 UTC m=+3451.238746130" watchObservedRunningTime="2025-12-01 20:36:55.954298566 +0000 UTC m=+3451.241790235" Dec 01 20:37:01 crc kubenswrapper[4960]: I1201 20:37:01.109815 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:01 crc kubenswrapper[4960]: I1201 20:37:01.110395 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:01 crc kubenswrapper[4960]: I1201 20:37:01.166921 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:02 crc kubenswrapper[4960]: I1201 20:37:02.054283 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:02 crc kubenswrapper[4960]: I1201 20:37:02.103332 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.032730 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h2c6d" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="registry-server" containerID="cri-o://bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b" gracePeriod=2 Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.843736 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.846518 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.862015 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.900627 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnlh5\" (UniqueName: \"kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.901059 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.901102 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:04 crc kubenswrapper[4960]: I1201 20:37:04.926947 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.002851 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content\") pod \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.002944 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj9gc\" (UniqueName: \"kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc\") pod \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.003043 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities\") pod \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\" (UID: \"33234aa0-f048-4e27-b8fb-9c351bb6b2cd\") " Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.003421 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnlh5\" (UniqueName: \"kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.003515 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.003546 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.004069 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.005242 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities" (OuterVolumeSpecName: "utilities") pod "33234aa0-f048-4e27-b8fb-9c351bb6b2cd" (UID: "33234aa0-f048-4e27-b8fb-9c351bb6b2cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.005479 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.016201 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc" (OuterVolumeSpecName: "kube-api-access-zj9gc") pod "33234aa0-f048-4e27-b8fb-9c351bb6b2cd" (UID: "33234aa0-f048-4e27-b8fb-9c351bb6b2cd"). InnerVolumeSpecName "kube-api-access-zj9gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.026952 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnlh5\" (UniqueName: \"kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5\") pod \"redhat-operators-lw92h\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.053041 4960 generic.go:334] "Generic (PLEG): container finished" podID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerID="bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b" exitCode=0 Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.053082 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerDied","Data":"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b"} Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.053195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h2c6d" event={"ID":"33234aa0-f048-4e27-b8fb-9c351bb6b2cd","Type":"ContainerDied","Data":"24b9278dbe5202479a7562eb8822f300d363d58c9eebb59a1380a87368e84db4"} Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.053213 4960 scope.go:117] "RemoveContainer" containerID="bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.053342 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h2c6d" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.061278 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33234aa0-f048-4e27-b8fb-9c351bb6b2cd" (UID: "33234aa0-f048-4e27-b8fb-9c351bb6b2cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.105855 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.105888 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.105898 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj9gc\" (UniqueName: \"kubernetes.io/projected/33234aa0-f048-4e27-b8fb-9c351bb6b2cd-kube-api-access-zj9gc\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.128131 4960 scope.go:117] "RemoveContainer" containerID="57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.164469 4960 scope.go:117] "RemoveContainer" containerID="5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.222514 4960 scope.go:117] "RemoveContainer" containerID="bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b" Dec 01 20:37:05 crc kubenswrapper[4960]: E1201 20:37:05.223031 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b\": container with ID starting with bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b not found: ID does not exist" containerID="bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.223080 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b"} err="failed to get container status \"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b\": rpc error: code = NotFound desc = could not find container \"bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b\": container with ID starting with bf58e66b2f3b6eafd780462b32c8368e86ff245c5644130dfb09f967f2e6b95b not found: ID does not exist" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.223100 4960 scope.go:117] "RemoveContainer" containerID="57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c" Dec 01 20:37:05 crc kubenswrapper[4960]: E1201 20:37:05.223332 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c\": container with ID starting with 57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c not found: ID does not exist" containerID="57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.223412 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c"} err="failed to get container status \"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c\": rpc error: code = NotFound desc = could not find container \"57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c\": container with ID starting with 57b5c2773b5c000ba2bf95ab4b1437d79da4a4e75492cf89d53b8a9fd2e8101c not found: ID does not exist" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.223427 4960 scope.go:117] "RemoveContainer" containerID="5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3" Dec 01 20:37:05 crc kubenswrapper[4960]: E1201 20:37:05.225425 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3\": container with ID starting with 5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3 not found: ID does not exist" containerID="5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.225547 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3"} err="failed to get container status \"5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3\": rpc error: code = NotFound desc = could not find container \"5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3\": container with ID starting with 5a00635e2f17a3f7dad7aaec126fc62271ebaf1f21192abeaca796178cc70ad3 not found: ID does not exist" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.247339 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.442181 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.462173 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h2c6d"] Dec 01 20:37:05 crc kubenswrapper[4960]: I1201 20:37:05.802687 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:06 crc kubenswrapper[4960]: I1201 20:37:06.063552 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerStarted","Data":"3c147eaf87960abe982c101ec32de85ddcaff2bbd040b82305005f1e01fa3ec2"} Dec 01 20:37:07 crc kubenswrapper[4960]: I1201 20:37:07.086258 4960 generic.go:334] "Generic (PLEG): container finished" podID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerID="50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3" exitCode=0 Dec 01 20:37:07 crc kubenswrapper[4960]: I1201 20:37:07.086433 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerDied","Data":"50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3"} Dec 01 20:37:07 crc kubenswrapper[4960]: I1201 20:37:07.337352 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" path="/var/lib/kubelet/pods/33234aa0-f048-4e27-b8fb-9c351bb6b2cd/volumes" Dec 01 20:37:09 crc kubenswrapper[4960]: I1201 20:37:09.108556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerStarted","Data":"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55"} Dec 01 20:37:11 crc kubenswrapper[4960]: I1201 20:37:11.141341 4960 generic.go:334] "Generic (PLEG): container finished" podID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerID="0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55" exitCode=0 Dec 01 20:37:11 crc kubenswrapper[4960]: I1201 20:37:11.141428 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerDied","Data":"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55"} Dec 01 20:37:12 crc kubenswrapper[4960]: I1201 20:37:12.153583 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerStarted","Data":"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24"} Dec 01 20:37:12 crc kubenswrapper[4960]: I1201 20:37:12.175952 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lw92h" podStartSLOduration=3.687304941 podStartE2EDuration="8.175935444s" podCreationTimestamp="2025-12-01 20:37:04 +0000 UTC" firstStartedPulling="2025-12-01 20:37:07.088770439 +0000 UTC m=+3462.376262098" lastFinishedPulling="2025-12-01 20:37:11.577400912 +0000 UTC m=+3466.864892601" observedRunningTime="2025-12-01 20:37:12.170914198 +0000 UTC m=+3467.458405867" watchObservedRunningTime="2025-12-01 20:37:12.175935444 +0000 UTC m=+3467.463427113" Dec 01 20:37:15 crc kubenswrapper[4960]: I1201 20:37:15.248393 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:15 crc kubenswrapper[4960]: I1201 20:37:15.249072 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:16 crc kubenswrapper[4960]: I1201 20:37:16.312635 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lw92h" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="registry-server" probeResult="failure" output=< Dec 01 20:37:16 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:37:16 crc kubenswrapper[4960]: > Dec 01 20:37:25 crc kubenswrapper[4960]: I1201 20:37:25.306524 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:25 crc kubenswrapper[4960]: I1201 20:37:25.360882 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:25 crc kubenswrapper[4960]: I1201 20:37:25.541392 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:27 crc kubenswrapper[4960]: I1201 20:37:27.307267 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lw92h" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="registry-server" containerID="cri-o://afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24" gracePeriod=2 Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.207795 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.318901 4960 generic.go:334] "Generic (PLEG): container finished" podID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerID="afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24" exitCode=0 Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.318953 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerDied","Data":"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24"} Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.318984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lw92h" event={"ID":"55d16ad9-c69c-41b4-a20e-e002265bb310","Type":"ContainerDied","Data":"3c147eaf87960abe982c101ec32de85ddcaff2bbd040b82305005f1e01fa3ec2"} Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.319004 4960 scope.go:117] "RemoveContainer" containerID="afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.319183 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lw92h" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.340963 4960 scope.go:117] "RemoveContainer" containerID="0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.359814 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content\") pod \"55d16ad9-c69c-41b4-a20e-e002265bb310\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.360131 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities\") pod \"55d16ad9-c69c-41b4-a20e-e002265bb310\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.360244 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnlh5\" (UniqueName: \"kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5\") pod \"55d16ad9-c69c-41b4-a20e-e002265bb310\" (UID: \"55d16ad9-c69c-41b4-a20e-e002265bb310\") " Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.365072 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities" (OuterVolumeSpecName: "utilities") pod "55d16ad9-c69c-41b4-a20e-e002265bb310" (UID: "55d16ad9-c69c-41b4-a20e-e002265bb310"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.370047 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5" (OuterVolumeSpecName: "kube-api-access-pnlh5") pod "55d16ad9-c69c-41b4-a20e-e002265bb310" (UID: "55d16ad9-c69c-41b4-a20e-e002265bb310"). InnerVolumeSpecName "kube-api-access-pnlh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.386351 4960 scope.go:117] "RemoveContainer" containerID="50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.456135 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55d16ad9-c69c-41b4-a20e-e002265bb310" (UID: "55d16ad9-c69c-41b4-a20e-e002265bb310"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.463996 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.464030 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnlh5\" (UniqueName: \"kubernetes.io/projected/55d16ad9-c69c-41b4-a20e-e002265bb310-kube-api-access-pnlh5\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.464043 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55d16ad9-c69c-41b4-a20e-e002265bb310-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.487338 4960 scope.go:117] "RemoveContainer" containerID="afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24" Dec 01 20:37:28 crc kubenswrapper[4960]: E1201 20:37:28.488047 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24\": container with ID starting with afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24 not found: ID does not exist" containerID="afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.488102 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24"} err="failed to get container status \"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24\": rpc error: code = NotFound desc = could not find container \"afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24\": container with ID starting with afa43fddb0ce48dea4a368ddc201e467dbd877ae11f9d9c6e0dc1806f7147a24 not found: ID does not exist" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.488151 4960 scope.go:117] "RemoveContainer" containerID="0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55" Dec 01 20:37:28 crc kubenswrapper[4960]: E1201 20:37:28.488777 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55\": container with ID starting with 0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55 not found: ID does not exist" containerID="0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.488812 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55"} err="failed to get container status \"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55\": rpc error: code = NotFound desc = could not find container \"0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55\": container with ID starting with 0d6dc317697b62e84402a0f9240cdbc86fd1f99944bb412cf09475809e102f55 not found: ID does not exist" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.488830 4960 scope.go:117] "RemoveContainer" containerID="50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3" Dec 01 20:37:28 crc kubenswrapper[4960]: E1201 20:37:28.489088 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3\": container with ID starting with 50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3 not found: ID does not exist" containerID="50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.489163 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3"} err="failed to get container status \"50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3\": rpc error: code = NotFound desc = could not find container \"50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3\": container with ID starting with 50491d470274fc6ecb608c8618d0a5638e3aee9292e96beb6b0237dbb5c4d9e3 not found: ID does not exist" Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.652187 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:28 crc kubenswrapper[4960]: I1201 20:37:28.662163 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lw92h"] Dec 01 20:37:29 crc kubenswrapper[4960]: I1201 20:37:29.341025 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" path="/var/lib/kubelet/pods/55d16ad9-c69c-41b4-a20e-e002265bb310/volumes" Dec 01 20:39:13 crc kubenswrapper[4960]: I1201 20:39:13.870722 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:39:13 crc kubenswrapper[4960]: I1201 20:39:13.871362 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:39:43 crc kubenswrapper[4960]: I1201 20:39:43.871099 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:39:43 crc kubenswrapper[4960]: I1201 20:39:43.871691 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:40:04 crc kubenswrapper[4960]: I1201 20:40:04.721737 4960 generic.go:334] "Generic (PLEG): container finished" podID="d50fcead-df5c-40a0-b60f-4c195869572d" containerID="7177c7e65162a79f0fc380fe26e43880d1e43e8cf3cdd9214fc528d19099b7de" exitCode=0 Dec 01 20:40:04 crc kubenswrapper[4960]: I1201 20:40:04.722152 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d50fcead-df5c-40a0-b60f-4c195869572d","Type":"ContainerDied","Data":"7177c7e65162a79f0fc380fe26e43880d1e43e8cf3cdd9214fc528d19099b7de"} Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.324077 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.450711 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451040 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451064 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451166 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451199 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451216 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwns4\" (UniqueName: \"kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451246 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451265 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.451321 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config\") pod \"d50fcead-df5c-40a0-b60f-4c195869572d\" (UID: \"d50fcead-df5c-40a0-b60f-4c195869572d\") " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.453516 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data" (OuterVolumeSpecName: "config-data") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.453851 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.459157 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.459389 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4" (OuterVolumeSpecName: "kube-api-access-kwns4") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "kube-api-access-kwns4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.484967 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.484991 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.491312 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.514108 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555039 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555081 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555118 4960 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555148 4960 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555161 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwns4\" (UniqueName: \"kubernetes.io/projected/d50fcead-df5c-40a0-b60f-4c195869572d-kube-api-access-kwns4\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555175 4960 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-config-data\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555187 4960 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d50fcead-df5c-40a0-b60f-4c195869572d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.555199 4960 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d50fcead-df5c-40a0-b60f-4c195869572d-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.591577 4960 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.657970 4960 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.741019 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d50fcead-df5c-40a0-b60f-4c195869572d","Type":"ContainerDied","Data":"e2061bcfd548ad6bb478f0ff42b7df6ec6326e0944ed282d67015aa7e26e72e0"} Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.741056 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2061bcfd548ad6bb478f0ff42b7df6ec6326e0944ed282d67015aa7e26e72e0" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.741092 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.928680 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d50fcead-df5c-40a0-b60f-4c195869572d" (UID: "d50fcead-df5c-40a0-b60f-4c195869572d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:40:06 crc kubenswrapper[4960]: I1201 20:40:06.964418 4960 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d50fcead-df5c-40a0-b60f-4c195869572d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 01 20:40:13 crc kubenswrapper[4960]: I1201 20:40:13.870934 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:40:13 crc kubenswrapper[4960]: I1201 20:40:13.871555 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:40:13 crc kubenswrapper[4960]: I1201 20:40:13.871612 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:40:13 crc kubenswrapper[4960]: I1201 20:40:13.872588 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:40:13 crc kubenswrapper[4960]: I1201 20:40:13.872663 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c" gracePeriod=600 Dec 01 20:40:14 crc kubenswrapper[4960]: I1201 20:40:14.823459 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c" exitCode=0 Dec 01 20:40:14 crc kubenswrapper[4960]: I1201 20:40:14.823556 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c"} Dec 01 20:40:14 crc kubenswrapper[4960]: I1201 20:40:14.824153 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed"} Dec 01 20:40:14 crc kubenswrapper[4960]: I1201 20:40:14.824182 4960 scope.go:117] "RemoveContainer" containerID="5bb28e762d39157c403712486db02ecc3df7b55f7bfdd3c5dfe76853f71fee4d" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.212749 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213243 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="extract-utilities" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213262 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="extract-utilities" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213281 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="extract-content" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213290 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="extract-content" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213320 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213329 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213344 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213349 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213359 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="extract-content" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213366 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="extract-content" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213382 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="extract-utilities" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213390 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="extract-utilities" Dec 01 20:40:15 crc kubenswrapper[4960]: E1201 20:40:15.213411 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d50fcead-df5c-40a0-b60f-4c195869572d" containerName="tempest-tests-tempest-tests-runner" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213419 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="d50fcead-df5c-40a0-b60f-4c195869572d" containerName="tempest-tests-tempest-tests-runner" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213674 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d16ad9-c69c-41b4-a20e-e002265bb310" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213708 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="33234aa0-f048-4e27-b8fb-9c351bb6b2cd" containerName="registry-server" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.213722 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="d50fcead-df5c-40a0-b60f-4c195869572d" containerName="tempest-tests-tempest-tests-runner" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.214551 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.216945 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-r2cdg" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.225418 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.340665 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mb9k\" (UniqueName: \"kubernetes.io/projected/468c1f0b-62e1-492a-8496-288dc865ea98-kube-api-access-9mb9k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.340954 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.443178 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mb9k\" (UniqueName: \"kubernetes.io/projected/468c1f0b-62e1-492a-8496-288dc865ea98-kube-api-access-9mb9k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.443695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.446791 4960 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.467332 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mb9k\" (UniqueName: \"kubernetes.io/projected/468c1f0b-62e1-492a-8496-288dc865ea98-kube-api-access-9mb9k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.496697 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"468c1f0b-62e1-492a-8496-288dc865ea98\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.538184 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.984902 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 01 20:40:15 crc kubenswrapper[4960]: W1201 20:40:15.993167 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod468c1f0b_62e1_492a_8496_288dc865ea98.slice/crio-39cb19f6a399ded69b91a8b664d9dc9affbf9fc1d413bd1812abb209b9043ac2 WatchSource:0}: Error finding container 39cb19f6a399ded69b91a8b664d9dc9affbf9fc1d413bd1812abb209b9043ac2: Status 404 returned error can't find the container with id 39cb19f6a399ded69b91a8b664d9dc9affbf9fc1d413bd1812abb209b9043ac2 Dec 01 20:40:15 crc kubenswrapper[4960]: I1201 20:40:15.996508 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:40:16 crc kubenswrapper[4960]: I1201 20:40:16.846055 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"468c1f0b-62e1-492a-8496-288dc865ea98","Type":"ContainerStarted","Data":"39cb19f6a399ded69b91a8b664d9dc9affbf9fc1d413bd1812abb209b9043ac2"} Dec 01 20:40:18 crc kubenswrapper[4960]: I1201 20:40:18.870437 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"468c1f0b-62e1-492a-8496-288dc865ea98","Type":"ContainerStarted","Data":"9393b4b285c1fc3de529cf397ab8b76d903d426475fbd1b97db02d53fbb5280c"} Dec 01 20:40:18 crc kubenswrapper[4960]: I1201 20:40:18.895100 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.844425764 podStartE2EDuration="3.894712412s" podCreationTimestamp="2025-12-01 20:40:15 +0000 UTC" firstStartedPulling="2025-12-01 20:40:15.996162044 +0000 UTC m=+3651.283653713" lastFinishedPulling="2025-12-01 20:40:18.046448692 +0000 UTC m=+3653.333940361" observedRunningTime="2025-12-01 20:40:18.88984664 +0000 UTC m=+3654.177338309" watchObservedRunningTime="2025-12-01 20:40:18.894712412 +0000 UTC m=+3654.182204101" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.535234 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-khtrz/must-gather-zq59t"] Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.537437 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.549240 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-khtrz"/"default-dockercfg-hmbgm" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.575375 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-khtrz"/"kube-root-ca.crt" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.576172 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-khtrz"/"openshift-service-ca.crt" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.581964 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-khtrz/must-gather-zq59t"] Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.712011 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68tfr\" (UniqueName: \"kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.712127 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.829316 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.829491 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68tfr\" (UniqueName: \"kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.830302 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:42 crc kubenswrapper[4960]: I1201 20:40:42.864086 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68tfr\" (UniqueName: \"kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr\") pod \"must-gather-zq59t\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:43 crc kubenswrapper[4960]: I1201 20:40:43.154656 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:40:43 crc kubenswrapper[4960]: I1201 20:40:43.726462 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-khtrz/must-gather-zq59t"] Dec 01 20:40:44 crc kubenswrapper[4960]: I1201 20:40:44.116578 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/must-gather-zq59t" event={"ID":"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e","Type":"ContainerStarted","Data":"1159e89c52845dc8515701dc7ff56d8d07fb2b745e9b93ba2b195f28f21234a2"} Dec 01 20:40:50 crc kubenswrapper[4960]: I1201 20:40:50.177665 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/must-gather-zq59t" event={"ID":"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e","Type":"ContainerStarted","Data":"372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0"} Dec 01 20:40:50 crc kubenswrapper[4960]: I1201 20:40:50.178289 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/must-gather-zq59t" event={"ID":"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e","Type":"ContainerStarted","Data":"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e"} Dec 01 20:40:50 crc kubenswrapper[4960]: I1201 20:40:50.196970 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-khtrz/must-gather-zq59t" podStartSLOduration=2.778000091 podStartE2EDuration="8.196949698s" podCreationTimestamp="2025-12-01 20:40:42 +0000 UTC" firstStartedPulling="2025-12-01 20:40:43.722577929 +0000 UTC m=+3679.010069598" lastFinishedPulling="2025-12-01 20:40:49.141527536 +0000 UTC m=+3684.429019205" observedRunningTime="2025-12-01 20:40:50.194716219 +0000 UTC m=+3685.482207888" watchObservedRunningTime="2025-12-01 20:40:50.196949698 +0000 UTC m=+3685.484441367" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.553304 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-khtrz/crc-debug-vbh4r"] Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.555580 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.653799 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgc6g\" (UniqueName: \"kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.653916 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.755168 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgc6g\" (UniqueName: \"kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.755255 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.755368 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.786044 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgc6g\" (UniqueName: \"kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g\") pod \"crc-debug-vbh4r\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:53 crc kubenswrapper[4960]: I1201 20:40:53.877804 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:40:54 crc kubenswrapper[4960]: I1201 20:40:54.235841 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" event={"ID":"e147cf15-909f-400c-bf49-07d8468a7196","Type":"ContainerStarted","Data":"805c39eb00f8a757c2a6a66668d73293f8ade6f8b11936486b931f56d98f2d28"} Dec 01 20:41:08 crc kubenswrapper[4960]: I1201 20:41:08.392616 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" event={"ID":"e147cf15-909f-400c-bf49-07d8468a7196","Type":"ContainerStarted","Data":"8ad2581999e3061026c88f9d9e4c3452b7ac4ea7925868cbd0dd9869cf37be6b"} Dec 01 20:41:08 crc kubenswrapper[4960]: I1201 20:41:08.416887 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" podStartSLOduration=2.131700816 podStartE2EDuration="15.416870306s" podCreationTimestamp="2025-12-01 20:40:53 +0000 UTC" firstStartedPulling="2025-12-01 20:40:53.928588394 +0000 UTC m=+3689.216080073" lastFinishedPulling="2025-12-01 20:41:07.213757894 +0000 UTC m=+3702.501249563" observedRunningTime="2025-12-01 20:41:08.406523354 +0000 UTC m=+3703.694015033" watchObservedRunningTime="2025-12-01 20:41:08.416870306 +0000 UTC m=+3703.704361975" Dec 01 20:42:01 crc kubenswrapper[4960]: I1201 20:42:01.931912 4960 generic.go:334] "Generic (PLEG): container finished" podID="e147cf15-909f-400c-bf49-07d8468a7196" containerID="8ad2581999e3061026c88f9d9e4c3452b7ac4ea7925868cbd0dd9869cf37be6b" exitCode=0 Dec 01 20:42:01 crc kubenswrapper[4960]: I1201 20:42:01.931992 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" event={"ID":"e147cf15-909f-400c-bf49-07d8468a7196","Type":"ContainerDied","Data":"8ad2581999e3061026c88f9d9e4c3452b7ac4ea7925868cbd0dd9869cf37be6b"} Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.068013 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.109382 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-vbh4r"] Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.120291 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-vbh4r"] Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.196051 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host\") pod \"e147cf15-909f-400c-bf49-07d8468a7196\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.196139 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgc6g\" (UniqueName: \"kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g\") pod \"e147cf15-909f-400c-bf49-07d8468a7196\" (UID: \"e147cf15-909f-400c-bf49-07d8468a7196\") " Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.196214 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host" (OuterVolumeSpecName: "host") pod "e147cf15-909f-400c-bf49-07d8468a7196" (UID: "e147cf15-909f-400c-bf49-07d8468a7196"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.196872 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e147cf15-909f-400c-bf49-07d8468a7196-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.204196 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g" (OuterVolumeSpecName: "kube-api-access-bgc6g") pod "e147cf15-909f-400c-bf49-07d8468a7196" (UID: "e147cf15-909f-400c-bf49-07d8468a7196"). InnerVolumeSpecName "kube-api-access-bgc6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.299668 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgc6g\" (UniqueName: \"kubernetes.io/projected/e147cf15-909f-400c-bf49-07d8468a7196-kube-api-access-bgc6g\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.335828 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e147cf15-909f-400c-bf49-07d8468a7196" path="/var/lib/kubelet/pods/e147cf15-909f-400c-bf49-07d8468a7196/volumes" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.954553 4960 scope.go:117] "RemoveContainer" containerID="8ad2581999e3061026c88f9d9e4c3452b7ac4ea7925868cbd0dd9869cf37be6b" Dec 01 20:42:03 crc kubenswrapper[4960]: I1201 20:42:03.954588 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-vbh4r" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.310642 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-khtrz/crc-debug-pxb9k"] Dec 01 20:42:04 crc kubenswrapper[4960]: E1201 20:42:04.311190 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e147cf15-909f-400c-bf49-07d8468a7196" containerName="container-00" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.311204 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="e147cf15-909f-400c-bf49-07d8468a7196" containerName="container-00" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.311425 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="e147cf15-909f-400c-bf49-07d8468a7196" containerName="container-00" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.312225 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.424456 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.424937 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2rsm\" (UniqueName: \"kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.526337 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2rsm\" (UniqueName: \"kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.526507 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.526643 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.549047 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2rsm\" (UniqueName: \"kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm\") pod \"crc-debug-pxb9k\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.632997 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.969564 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" event={"ID":"34d09d1f-00e3-49e3-bf09-11eb498ebde4","Type":"ContainerStarted","Data":"66e11cef59fdf77e0d32073563e6eebda0d26baa419e0fb8de75f903f74ee0ea"} Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.969902 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" event={"ID":"34d09d1f-00e3-49e3-bf09-11eb498ebde4","Type":"ContainerStarted","Data":"969e3f356b17411395db5fd4a0229acb7621c84d6f5e7ff6177b4a084458ac7d"} Dec 01 20:42:04 crc kubenswrapper[4960]: I1201 20:42:04.986493 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" podStartSLOduration=0.986467875 podStartE2EDuration="986.467875ms" podCreationTimestamp="2025-12-01 20:42:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:42:04.983043768 +0000 UTC m=+3760.270535477" watchObservedRunningTime="2025-12-01 20:42:04.986467875 +0000 UTC m=+3760.273959554" Dec 01 20:42:05 crc kubenswrapper[4960]: I1201 20:42:05.981413 4960 generic.go:334] "Generic (PLEG): container finished" podID="34d09d1f-00e3-49e3-bf09-11eb498ebde4" containerID="66e11cef59fdf77e0d32073563e6eebda0d26baa419e0fb8de75f903f74ee0ea" exitCode=0 Dec 01 20:42:05 crc kubenswrapper[4960]: I1201 20:42:05.981463 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" event={"ID":"34d09d1f-00e3-49e3-bf09-11eb498ebde4","Type":"ContainerDied","Data":"66e11cef59fdf77e0d32073563e6eebda0d26baa419e0fb8de75f903f74ee0ea"} Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.111144 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.155348 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-pxb9k"] Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.166325 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-pxb9k"] Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.283227 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host\") pod \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.283402 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host" (OuterVolumeSpecName: "host") pod "34d09d1f-00e3-49e3-bf09-11eb498ebde4" (UID: "34d09d1f-00e3-49e3-bf09-11eb498ebde4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.283481 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2rsm\" (UniqueName: \"kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm\") pod \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\" (UID: \"34d09d1f-00e3-49e3-bf09-11eb498ebde4\") " Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.283948 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/34d09d1f-00e3-49e3-bf09-11eb498ebde4-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.292430 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm" (OuterVolumeSpecName: "kube-api-access-n2rsm") pod "34d09d1f-00e3-49e3-bf09-11eb498ebde4" (UID: "34d09d1f-00e3-49e3-bf09-11eb498ebde4"). InnerVolumeSpecName "kube-api-access-n2rsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.337816 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34d09d1f-00e3-49e3-bf09-11eb498ebde4" path="/var/lib/kubelet/pods/34d09d1f-00e3-49e3-bf09-11eb498ebde4/volumes" Dec 01 20:42:07 crc kubenswrapper[4960]: I1201 20:42:07.385903 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2rsm\" (UniqueName: \"kubernetes.io/projected/34d09d1f-00e3-49e3-bf09-11eb498ebde4-kube-api-access-n2rsm\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.002499 4960 scope.go:117] "RemoveContainer" containerID="66e11cef59fdf77e0d32073563e6eebda0d26baa419e0fb8de75f903f74ee0ea" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.002557 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-pxb9k" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.330271 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-khtrz/crc-debug-rnk99"] Dec 01 20:42:08 crc kubenswrapper[4960]: E1201 20:42:08.330790 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d09d1f-00e3-49e3-bf09-11eb498ebde4" containerName="container-00" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.330806 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d09d1f-00e3-49e3-bf09-11eb498ebde4" containerName="container-00" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.331108 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="34d09d1f-00e3-49e3-bf09-11eb498ebde4" containerName="container-00" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.332012 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.406902 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.406989 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxk2\" (UniqueName: \"kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.509651 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.509719 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxk2\" (UniqueName: \"kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.509849 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.539250 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxk2\" (UniqueName: \"kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2\") pod \"crc-debug-rnk99\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:08 crc kubenswrapper[4960]: I1201 20:42:08.650715 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:09 crc kubenswrapper[4960]: I1201 20:42:09.012557 4960 generic.go:334] "Generic (PLEG): container finished" podID="4f1ca217-35f2-4854-9aa6-9fa70268b8ac" containerID="ea98265305753bcd7b1bc8ef3d52e21ce5b2637dbf4076b39be83e4fc0b8fc94" exitCode=0 Dec 01 20:42:09 crc kubenswrapper[4960]: I1201 20:42:09.012599 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-rnk99" event={"ID":"4f1ca217-35f2-4854-9aa6-9fa70268b8ac","Type":"ContainerDied","Data":"ea98265305753bcd7b1bc8ef3d52e21ce5b2637dbf4076b39be83e4fc0b8fc94"} Dec 01 20:42:09 crc kubenswrapper[4960]: I1201 20:42:09.012879 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/crc-debug-rnk99" event={"ID":"4f1ca217-35f2-4854-9aa6-9fa70268b8ac","Type":"ContainerStarted","Data":"3c7cb699de15d7091083c04969cbe6361d3dc7f01919d993db3c5646fc33687d"} Dec 01 20:42:09 crc kubenswrapper[4960]: I1201 20:42:09.082004 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-rnk99"] Dec 01 20:42:09 crc kubenswrapper[4960]: I1201 20:42:09.093779 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-khtrz/crc-debug-rnk99"] Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.128396 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.241391 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrxk2\" (UniqueName: \"kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2\") pod \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.241486 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host\") pod \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\" (UID: \"4f1ca217-35f2-4854-9aa6-9fa70268b8ac\") " Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.241639 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host" (OuterVolumeSpecName: "host") pod "4f1ca217-35f2-4854-9aa6-9fa70268b8ac" (UID: "4f1ca217-35f2-4854-9aa6-9fa70268b8ac"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.241928 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.247379 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2" (OuterVolumeSpecName: "kube-api-access-lrxk2") pod "4f1ca217-35f2-4854-9aa6-9fa70268b8ac" (UID: "4f1ca217-35f2-4854-9aa6-9fa70268b8ac"). InnerVolumeSpecName "kube-api-access-lrxk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:42:10 crc kubenswrapper[4960]: I1201 20:42:10.343762 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrxk2\" (UniqueName: \"kubernetes.io/projected/4f1ca217-35f2-4854-9aa6-9fa70268b8ac-kube-api-access-lrxk2\") on node \"crc\" DevicePath \"\"" Dec 01 20:42:11 crc kubenswrapper[4960]: I1201 20:42:11.031276 4960 scope.go:117] "RemoveContainer" containerID="ea98265305753bcd7b1bc8ef3d52e21ce5b2637dbf4076b39be83e4fc0b8fc94" Dec 01 20:42:11 crc kubenswrapper[4960]: I1201 20:42:11.031305 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/crc-debug-rnk99" Dec 01 20:42:11 crc kubenswrapper[4960]: I1201 20:42:11.339060 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1ca217-35f2-4854-9aa6-9fa70268b8ac" path="/var/lib/kubelet/pods/4f1ca217-35f2-4854-9aa6-9fa70268b8ac/volumes" Dec 01 20:42:36 crc kubenswrapper[4960]: I1201 20:42:36.827036 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/init-config-reloader/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.068926 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/alertmanager/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.087790 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/init-config-reloader/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.111852 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/config-reloader/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.258186 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d89589d9d-9dglh_9c816e6d-aa4e-46cc-8494-39d3f334a028/barbican-api/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.269549 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d89589d9d-9dglh_9c816e6d-aa4e-46cc-8494-39d3f334a028/barbican-api-log/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.316917 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f67669b46-khzh6_4771a4c0-c72c-4819-953e-ed6ecd9d1ef6/barbican-keystone-listener/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.512868 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f647bc89f-mwqdp_c6e92e0f-eb53-4d9a-af6a-4294f774a870/barbican-worker/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.513929 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f67669b46-khzh6_4771a4c0-c72c-4819-953e-ed6ecd9d1ef6/barbican-keystone-listener-log/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.553685 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f647bc89f-mwqdp_c6e92e0f-eb53-4d9a-af6a-4294f774a870/barbican-worker-log/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.779486 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb_098a01da-e6f9-4bc8-8249-72bbeaf949f7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:37 crc kubenswrapper[4960]: I1201 20:42:37.895498 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/ceilometer-central-agent/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.094193 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/proxy-httpd/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.109448 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/ceilometer-notification-agent/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.162970 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/sg-core/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.336574 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b45703c2-5b71-47a4-835a-cee374c228b7/cinder-api/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.337907 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b45703c2-5b71-47a4-835a-cee374c228b7/cinder-api-log/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.479811 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c0bee6fd-952d-4a33-83dc-c93625f15979/cinder-scheduler/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.543278 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c0bee6fd-952d-4a33-83dc-c93625f15979/probe/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.683225 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf/cloudkitty-api/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.744585 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf/cloudkitty-api-log/0.log" Dec 01 20:42:38 crc kubenswrapper[4960]: I1201 20:42:38.857211 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_8e8a9388-578d-4d41-a2cf-d2ad6e344288/loki-compactor/0.log" Dec 01 20:42:39 crc kubenswrapper[4960]: I1201 20:42:39.017439 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-56cd74f89f-jxqfj_2e566703-0d27-4306-bd42-7a91da29c736/loki-distributor/0.log" Dec 01 20:42:39 crc kubenswrapper[4960]: I1201 20:42:39.081370 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-mpddm_5cc4eeaa-88b7-4439-a39f-5e889845181e/gateway/0.log" Dec 01 20:42:39 crc kubenswrapper[4960]: I1201 20:42:39.251291 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-xw4bs_a8b0ac50-f375-427e-b103-0bfd272201ea/gateway/0.log" Dec 01 20:42:39 crc kubenswrapper[4960]: I1201 20:42:39.296638 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38/loki-index-gateway/0.log" Dec 01 20:42:39 crc kubenswrapper[4960]: I1201 20:42:39.807911 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-779849886d-hf7g4_2d2cf081-f13a-411b-bf24-5302a947dd9a/loki-query-frontend/0.log" Dec 01 20:42:40 crc kubenswrapper[4960]: I1201 20:42:40.035408 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_969241f0-8883-4570-94d4-fdb20ee094cd/loki-ingester/0.log" Dec 01 20:42:40 crc kubenswrapper[4960]: I1201 20:42:40.339274 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt_b6b5eb9b-033d-4417-9289-953b2933520e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:40 crc kubenswrapper[4960]: I1201 20:42:40.619327 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2_f45d70f4-68fe-491f-aa15-4e7d6a0a7812/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:40 crc kubenswrapper[4960]: I1201 20:42:40.653937 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-548665d79b-gd6lz_60517c3e-f430-453c-b6d4-3cdf124e100b/loki-querier/0.log" Dec 01 20:42:40 crc kubenswrapper[4960]: I1201 20:42:40.817037 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/init/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.160287 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/init/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.160422 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd_476f3fb9-9214-4448-9ff4-a5489883199b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.181650 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/dnsmasq-dns/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.405942 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a341d11-27da-4def-a478-c878a69e2e6b/glance-log/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.438197 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a341d11-27da-4def-a478-c878a69e2e6b/glance-httpd/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.570082 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1f3729c-7da0-4c16-b370-cacabf55f2b7/glance-httpd/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.613518 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1f3729c-7da0-4c16-b370-cacabf55f2b7/glance-log/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.703243 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr_cc7c39de-f6ea-4f00-ae86-a8f74ec35b71/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:41 crc kubenswrapper[4960]: I1201 20:42:41.937027 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9bb2x_4aa6823d-0480-490a-977b-801110a10181/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.205079 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e118a262-1b53-4d2a-adce-3802b626cff1/kube-state-metrics/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.415105 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q_4c5eadde-8ce4-4009-9128-66d4d9657dfc/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.445503 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5f8c78b6fc-fdfln_987ef811-8e23-491e-98ff-64fcacab0f37/keystone-api/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.499024 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a/cloudkitty-proc/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.790096 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd769db7-hsw86_aee39569-b67b-419a-ad6a-461bbed0d585/neutron-httpd/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.886103 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd769db7-hsw86_aee39569-b67b-419a-ad6a-461bbed0d585/neutron-api/0.log" Dec 01 20:42:42 crc kubenswrapper[4960]: I1201 20:42:42.901849 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58_6e32c90f-f0de-404d-be43-6953814aa12a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.385886 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_878b6439-3b49-40d4-80e2-f01694590b67/nova-cell0-conductor-conductor/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.502786 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3cb295cb-a487-406c-8f1d-517de7c1d245/nova-api-log/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.659011 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3cb295cb-a487-406c-8f1d-517de7c1d245/nova-api-api/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.747915 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_38ad16c2-10a9-4a20-a070-e1ed80f62684/nova-cell1-conductor-conductor/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.816784 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_35daed02-01d5-4757-aa71-9c38e1397d19/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.871521 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.871575 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:42:43 crc kubenswrapper[4960]: I1201 20:42:43.971947 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-stff5_e37fd55a-a5e3-4844-93b3-a22664d77bcd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.175073 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_feac0db7-ea9a-4add-9c01-f942bc15b1bf/nova-metadata-log/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.320668 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_08cc519f-2c0d-444a-845a-9f403c9101be/nova-scheduler-scheduler/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.621823 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/mysql-bootstrap/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.813986 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/mysql-bootstrap/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.827515 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/galera/0.log" Dec 01 20:42:44 crc kubenswrapper[4960]: I1201 20:42:44.990137 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/mysql-bootstrap/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.238441 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/mysql-bootstrap/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.253043 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/galera/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.293568 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_feac0db7-ea9a-4add-9c01-f942bc15b1bf/nova-metadata-metadata/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.416894 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_5f778c19-cf9c-4c12-811d-9bc33aae5e90/openstackclient/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.552419 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-trsdq_5c1ff07c-096b-4952-a0cb-32fdcfaf154a/openstack-network-exporter/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.668128 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nrf7z_87e16436-45f5-4392-b524-b86f1cc5c0b2/ovn-controller/0.log" Dec 01 20:42:45 crc kubenswrapper[4960]: I1201 20:42:45.805956 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server-init/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.019914 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.029267 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server-init/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.046386 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovs-vswitchd/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.261050 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_07787692-57cc-473c-ab9b-276004fb9220/openstack-network-exporter/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.286363 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-bhl97_95e8c909-a7d7-4461-8fc7-c7644ad41000/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.376843 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_07787692-57cc-473c-ab9b-276004fb9220/ovn-northd/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.489771 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a7eb168-4c2f-46c0-9292-de8e5c4383a5/ovsdbserver-nb/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.512692 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a7eb168-4c2f-46c0-9292-de8e5c4383a5/openstack-network-exporter/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.766257 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1c6f94f2-a256-4e77-b6d1-39706b7c6ddc/openstack-network-exporter/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.777664 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1c6f94f2-a256-4e77-b6d1-39706b7c6ddc/ovsdbserver-sb/0.log" Dec 01 20:42:46 crc kubenswrapper[4960]: I1201 20:42:46.991399 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c7bf679b-vg6f9_f75be90e-8df9-4f2a-bbdb-6202c7b56cf9/placement-api/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.088799 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c7bf679b-vg6f9_f75be90e-8df9-4f2a-bbdb-6202c7b56cf9/placement-log/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.133631 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/init-config-reloader/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.247946 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/init-config-reloader/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.256556 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/config-reloader/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.354481 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/prometheus/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.369674 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/thanos-sidecar/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.482333 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/setup-container/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.753224 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/setup-container/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.778057 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/setup-container/0.log" Dec 01 20:42:47 crc kubenswrapper[4960]: I1201 20:42:47.834415 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/rabbitmq/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.058825 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc_1d22d084-b469-42e1-afdf-6c8ba0c4c59e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.062084 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/rabbitmq/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.138625 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/setup-container/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.297073 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ndr5m_03306a2e-8c24-4479-bf7c-f6fbae3dd8ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.300694 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n_bca07824-2e7a-4578-b833-02232c52ca0c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.535350 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-t8vj8_91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.607953 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mmbqh_9ab87b90-6085-4c16-a656-a6775aeb0b68/ssh-known-hosts-edpm-deployment/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.876888 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696549d5f9-6rmsc_c99a2254-ef5a-47fd-8ed6-bc98e9671077/proxy-server/0.log" Dec 01 20:42:48 crc kubenswrapper[4960]: I1201 20:42:48.970019 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696549d5f9-6rmsc_c99a2254-ef5a-47fd-8ed6-bc98e9671077/proxy-httpd/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.029594 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gmqjb_aa3614b7-1a8b-4007-9e01-f6340ff9b0fe/swift-ring-rebalance/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.098530 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-auditor/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.264256 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-reaper/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.332748 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-replicator/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.380942 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-server/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.416391 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-auditor/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.539195 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-replicator/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.585460 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-server/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.637568 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-updater/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.672203 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-auditor/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.785653 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-expirer/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.845044 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-replicator/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.931352 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-server/0.log" Dec 01 20:42:49 crc kubenswrapper[4960]: I1201 20:42:49.945865 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-updater/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.069844 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/rsync/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.101722 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/swift-recon-cron/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.285365 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc_00d8a61e-f3a9-4f45-9ff3-a196f814de12/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.349161 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d50fcead-df5c-40a0-b60f-4c195869572d/tempest-tests-tempest-tests-runner/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.527769 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_468c1f0b-62e1-492a-8496-288dc865ea98/test-operator-logs-container/0.log" Dec 01 20:42:50 crc kubenswrapper[4960]: I1201 20:42:50.635828 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8_d8dcd369-d6cb-4472-8530-9c4d58ac2a0c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:42:55 crc kubenswrapper[4960]: I1201 20:42:55.949497 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6befffd7-b8c4-4c6c-900c-f6e1728f2a9e/memcached/0.log" Dec 01 20:43:13 crc kubenswrapper[4960]: I1201 20:43:13.871290 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:43:13 crc kubenswrapper[4960]: I1201 20:43:13.871808 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:43:16 crc kubenswrapper[4960]: I1201 20:43:16.631528 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:43:16 crc kubenswrapper[4960]: I1201 20:43:16.815105 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:43:16 crc kubenswrapper[4960]: I1201 20:43:16.851153 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:43:16 crc kubenswrapper[4960]: I1201 20:43:16.875434 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.255801 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.294433 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.306660 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/extract/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.480160 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8l7sf_238b3987-76a6-401d-a33c-5c2efd1bafcf/kube-rbac-proxy/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.582922 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8l7sf_238b3987-76a6-401d-a33c-5c2efd1bafcf/manager/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.597643 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-p5qfh_cce3a29f-5414-4fee-a27c-9b6286930903/kube-rbac-proxy/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.757887 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-p5qfh_cce3a29f-5414-4fee-a27c-9b6286930903/manager/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.841501 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6tbkt_a51010d5-7158-47ce-9428-25f5b2ca871f/kube-rbac-proxy/0.log" Dec 01 20:43:17 crc kubenswrapper[4960]: I1201 20:43:17.862499 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6tbkt_a51010d5-7158-47ce-9428-25f5b2ca871f/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.060591 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-pbxhm_018cf742-9a47-4355-879a-9e064f358748/kube-rbac-proxy/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.146489 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-pbxhm_018cf742-9a47-4355-879a-9e064f358748/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.231609 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-6mqj7_0a07e8af-0ea2-4b3e-b9b7-b01f88025904/kube-rbac-proxy/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.267604 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-6mqj7_0a07e8af-0ea2-4b3e-b9b7-b01f88025904/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.370070 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bsfbn_dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d/kube-rbac-proxy/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.439198 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bsfbn_dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.575641 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pnrhx_b801b659-3926-4a9f-a36f-f752a6bd9563/kube-rbac-proxy/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.727680 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pnrhx_b801b659-3926-4a9f-a36f-f752a6bd9563/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.798109 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zggh7_58815c45-1d27-4f86-9b6a-ffbb04c4e250/kube-rbac-proxy/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.803041 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zggh7_58815c45-1d27-4f86-9b6a-ffbb04c4e250/manager/0.log" Dec 01 20:43:18 crc kubenswrapper[4960]: I1201 20:43:18.962262 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-sfsvg_8b922299-0b5e-4595-bdab-1c759fd99465/kube-rbac-proxy/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.054566 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-sfsvg_8b922299-0b5e-4595-bdab-1c759fd99465/manager/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.172739 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-hcpbb_3b481993-f9fd-4d99-98e3-04b51c186e20/kube-rbac-proxy/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.177105 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-hcpbb_3b481993-f9fd-4d99-98e3-04b51c186e20/manager/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.365540 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-q8bv7_8c0036c6-9f32-4163-866d-c1f5786c5a2e/kube-rbac-proxy/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.463256 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-q8bv7_8c0036c6-9f32-4163-866d-c1f5786c5a2e/manager/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.574966 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-m24v2_de2501b9-bdd7-46c3-9128-20b2d9d5ab88/manager/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.583092 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-m24v2_de2501b9-bdd7-46c3-9128-20b2d9d5ab88/kube-rbac-proxy/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.739586 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-ld68p_0106714e-857b-4ade-9ee2-66fe8c12164c/kube-rbac-proxy/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.830149 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-ld68p_0106714e-857b-4ade-9ee2-66fe8c12164c/manager/0.log" Dec 01 20:43:19 crc kubenswrapper[4960]: I1201 20:43:19.932826 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cjgcs_b869ba57-9812-4684-8cbb-48fc606c6aec/kube-rbac-proxy/0.log" Dec 01 20:43:20 crc kubenswrapper[4960]: I1201 20:43:20.007667 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cjgcs_b869ba57-9812-4684-8cbb-48fc606c6aec/manager/0.log" Dec 01 20:43:20 crc kubenswrapper[4960]: I1201 20:43:20.120390 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6_6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c/manager/0.log" Dec 01 20:43:20 crc kubenswrapper[4960]: I1201 20:43:20.152904 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6_6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c/kube-rbac-proxy/0.log" Dec 01 20:43:20 crc kubenswrapper[4960]: I1201 20:43:20.492585 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8455849bfc-vgkq8_9179ddbd-9397-405d-93d5-e32ce4963305/operator/0.log" Dec 01 20:43:20 crc kubenswrapper[4960]: I1201 20:43:20.592002 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2jktp_d683e2dd-4bba-4833-ae18-4636090ef1e5/registry-server/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.077679 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-m9qpq_9b79fea5-fb83-4804-987d-b61e11999829/kube-rbac-proxy/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.239652 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-m9qpq_9b79fea5-fb83-4804-987d-b61e11999829/manager/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.249744 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vjb5b_b6018c51-4633-48dd-97ea-0d3d405f4e14/kube-rbac-proxy/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.300186 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vjb5b_b6018c51-4633-48dd-97ea-0d3d405f4e14/manager/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.481999 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5j8pz_123c59a4-e74e-4b51-9477-84d97ba60e36/operator/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.543924 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6zgq8_406103b8-0898-4d8f-badd-5f32648ad242/kube-rbac-proxy/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.620665 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8466c66f58-t6nmf_d52ae874-2bed-4c1a-9c8c-a3e140971d4a/manager/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.687470 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6zgq8_406103b8-0898-4d8f-badd-5f32648ad242/manager/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.763035 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-75c4b445dc-tfgqn_0b693fe0-7de6-4a0c-a931-21174d92c184/kube-rbac-proxy/0.log" Dec 01 20:43:21 crc kubenswrapper[4960]: I1201 20:43:21.977307 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rjb6d_2c77adc3-2d13-4c70-aefa-e4c238e61547/kube-rbac-proxy/0.log" Dec 01 20:43:22 crc kubenswrapper[4960]: I1201 20:43:22.060473 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rjb6d_2c77adc3-2d13-4c70-aefa-e4c238e61547/manager/0.log" Dec 01 20:43:22 crc kubenswrapper[4960]: I1201 20:43:22.130940 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-75c4b445dc-tfgqn_0b693fe0-7de6-4a0c-a931-21174d92c184/manager/0.log" Dec 01 20:43:22 crc kubenswrapper[4960]: I1201 20:43:22.156134 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k2567_82c72555-ba21-47f1-bfcd-ef954e788eb5/kube-rbac-proxy/0.log" Dec 01 20:43:22 crc kubenswrapper[4960]: I1201 20:43:22.177822 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k2567_82c72555-ba21-47f1-bfcd-ef954e788eb5/manager/0.log" Dec 01 20:43:40 crc kubenswrapper[4960]: I1201 20:43:40.833570 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9qhvh_952f446b-a8d2-4887-bc67-cddfb9e20791/control-plane-machine-set-operator/0.log" Dec 01 20:43:40 crc kubenswrapper[4960]: I1201 20:43:40.979137 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mhbt7_b18c64d6-cd4c-4d07-9591-4c9868b33b2c/kube-rbac-proxy/0.log" Dec 01 20:43:41 crc kubenswrapper[4960]: I1201 20:43:41.028033 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mhbt7_b18c64d6-cd4c-4d07-9591-4c9868b33b2c/machine-api-operator/0.log" Dec 01 20:43:43 crc kubenswrapper[4960]: I1201 20:43:43.871404 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:43:43 crc kubenswrapper[4960]: I1201 20:43:43.871935 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:43:43 crc kubenswrapper[4960]: I1201 20:43:43.871995 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:43:43 crc kubenswrapper[4960]: I1201 20:43:43.872870 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:43:43 crc kubenswrapper[4960]: I1201 20:43:43.872943 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" gracePeriod=600 Dec 01 20:43:43 crc kubenswrapper[4960]: E1201 20:43:43.993306 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:43:44 crc kubenswrapper[4960]: I1201 20:43:44.024177 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" exitCode=0 Dec 01 20:43:44 crc kubenswrapper[4960]: I1201 20:43:44.024219 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed"} Dec 01 20:43:44 crc kubenswrapper[4960]: I1201 20:43:44.024250 4960 scope.go:117] "RemoveContainer" containerID="26fe798e8ea15cb709526fceaaf8e93e930674fe3b77ae1d39eca33ffeb7962c" Dec 01 20:43:44 crc kubenswrapper[4960]: I1201 20:43:44.024978 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:43:44 crc kubenswrapper[4960]: E1201 20:43:44.025267 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:43:53 crc kubenswrapper[4960]: I1201 20:43:53.337319 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7dpzk_2e0d8acc-f304-45d4-b53c-dec0723c94e2/cert-manager-controller/0.log" Dec 01 20:43:53 crc kubenswrapper[4960]: I1201 20:43:53.517686 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-9lvxq_75151dc8-631d-42be-9b8c-e6b9a532c0da/cert-manager-cainjector/0.log" Dec 01 20:43:53 crc kubenswrapper[4960]: I1201 20:43:53.547324 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-fsw9v_12db57f7-a5f5-4a77-a412-d659e1490968/cert-manager-webhook/0.log" Dec 01 20:43:59 crc kubenswrapper[4960]: I1201 20:43:59.323813 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:43:59 crc kubenswrapper[4960]: E1201 20:43:59.324671 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:44:05 crc kubenswrapper[4960]: I1201 20:44:05.519182 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xpq6m_751bcd73-4c55-4fbb-a12d-0c9956a63e16/nmstate-console-plugin/0.log" Dec 01 20:44:05 crc kubenswrapper[4960]: I1201 20:44:05.694778 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-bwb2n_80d93577-3e89-47d7-b363-f22021509688/nmstate-handler/0.log" Dec 01 20:44:05 crc kubenswrapper[4960]: I1201 20:44:05.767756 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vh2zk_cba17f81-92f7-418b-a5c7-a184634c9acc/kube-rbac-proxy/0.log" Dec 01 20:44:05 crc kubenswrapper[4960]: I1201 20:44:05.851609 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vh2zk_cba17f81-92f7-418b-a5c7-a184634c9acc/nmstate-metrics/0.log" Dec 01 20:44:05 crc kubenswrapper[4960]: I1201 20:44:05.991598 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bvgrd_6377bee0-949f-4c1b-9758-c425ab4660bb/nmstate-operator/0.log" Dec 01 20:44:06 crc kubenswrapper[4960]: I1201 20:44:06.062322 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-5ndvb_138d040c-3866-4963-a2c8-26fb8a6c9811/nmstate-webhook/0.log" Dec 01 20:44:14 crc kubenswrapper[4960]: I1201 20:44:14.324338 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:44:14 crc kubenswrapper[4960]: E1201 20:44:14.325349 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:44:18 crc kubenswrapper[4960]: I1201 20:44:18.096311 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/kube-rbac-proxy/0.log" Dec 01 20:44:18 crc kubenswrapper[4960]: I1201 20:44:18.125857 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/manager/0.log" Dec 01 20:44:25 crc kubenswrapper[4960]: I1201 20:44:25.332268 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:44:25 crc kubenswrapper[4960]: E1201 20:44:25.333182 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:44:31 crc kubenswrapper[4960]: I1201 20:44:31.964015 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sm9mw_41410416-def6-44f6-9a25-14a6b9a50cec/kube-rbac-proxy/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.059100 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sm9mw_41410416-def6-44f6-9a25-14a6b9a50cec/controller/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.555612 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.762866 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.787158 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.803699 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:44:32 crc kubenswrapper[4960]: I1201 20:44:32.817565 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.040541 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.055445 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.071875 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.179316 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.406720 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.495698 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.748789 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.950541 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/controller/0.log" Dec 01 20:44:33 crc kubenswrapper[4960]: I1201 20:44:33.967394 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/frr-metrics/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.078625 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/kube-rbac-proxy/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.174495 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/kube-rbac-proxy-frr/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.223496 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/reloader/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.438385 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wjq2w_c7cef304-b739-45e6-b2c8-c729abfed45d/frr-k8s-webhook-server/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.523699 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f48564cc-pv444_8b72f468-b777-407c-8e36-30cc398644a6/manager/0.log" Dec 01 20:44:34 crc kubenswrapper[4960]: I1201 20:44:34.910409 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d7c4c664-7fttb_914a4b4c-d013-45ef-832d-35b600874887/webhook-server/0.log" Dec 01 20:44:35 crc kubenswrapper[4960]: I1201 20:44:35.155288 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/frr/0.log" Dec 01 20:44:35 crc kubenswrapper[4960]: I1201 20:44:35.522110 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qztxp_0c199599-aa8d-4e80-93f4-640a8ff141fa/kube-rbac-proxy/0.log" Dec 01 20:44:35 crc kubenswrapper[4960]: I1201 20:44:35.816961 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qztxp_0c199599-aa8d-4e80-93f4-640a8ff141fa/speaker/0.log" Dec 01 20:44:37 crc kubenswrapper[4960]: I1201 20:44:37.328393 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:44:37 crc kubenswrapper[4960]: E1201 20:44:37.328870 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:44:48 crc kubenswrapper[4960]: I1201 20:44:48.323914 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:44:48 crc kubenswrapper[4960]: E1201 20:44:48.324850 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:44:50 crc kubenswrapper[4960]: I1201 20:44:50.891890 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.108136 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.109187 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.186953 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.355548 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.365445 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/extract/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.388260 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.571169 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.748042 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.758684 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.781980 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.977377 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.978228 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:44:51 crc kubenswrapper[4960]: I1201 20:44:51.999248 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/extract/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.330316 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.566447 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.586145 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.608289 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.820297 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:44:52 crc kubenswrapper[4960]: I1201 20:44:52.838587 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.038446 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.239656 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/registry-server/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.308514 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.346499 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.354639 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.538801 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.607527 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.783208 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qsjbd_a968fcf3-9935-4cfe-abdc-82c62d9b3b12/marketplace-operator/0.log" Dec 01 20:44:53 crc kubenswrapper[4960]: I1201 20:44:53.908651 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.113749 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/registry-server/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.279837 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.340525 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.340689 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.410812 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.532747 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.695068 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.790226 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/registry-server/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.927561 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.939712 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:44:54 crc kubenswrapper[4960]: I1201 20:44:54.943729 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:44:55 crc kubenswrapper[4960]: I1201 20:44:55.100063 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:44:55 crc kubenswrapper[4960]: I1201 20:44:55.120688 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:44:55 crc kubenswrapper[4960]: I1201 20:44:55.595865 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/registry-server/0.log" Dec 01 20:44:59 crc kubenswrapper[4960]: I1201 20:44:59.324036 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:44:59 crc kubenswrapper[4960]: E1201 20:44:59.324771 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.211902 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj"] Dec 01 20:45:00 crc kubenswrapper[4960]: E1201 20:45:00.212711 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1ca217-35f2-4854-9aa6-9fa70268b8ac" containerName="container-00" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.212730 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1ca217-35f2-4854-9aa6-9fa70268b8ac" containerName="container-00" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.212996 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1ca217-35f2-4854-9aa6-9fa70268b8ac" containerName="container-00" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.213971 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.217673 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.217885 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.258087 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj"] Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.325478 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.325594 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zb82\" (UniqueName: \"kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.325640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.427279 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.427405 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zb82\" (UniqueName: \"kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.427435 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.428162 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.434721 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.444797 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zb82\" (UniqueName: \"kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82\") pod \"collect-profiles-29410365-8s8rj\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:00 crc kubenswrapper[4960]: I1201 20:45:00.550654 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:01 crc kubenswrapper[4960]: I1201 20:45:01.181340 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj"] Dec 01 20:45:01 crc kubenswrapper[4960]: I1201 20:45:01.782616 4960 generic.go:334] "Generic (PLEG): container finished" podID="fc634975-a76e-4aec-97e3-c1f1be39daa4" containerID="4a5df75a0c4e3900b8b82218d6a256eb7122441fe2b3211e8ec31dc71cc490ac" exitCode=0 Dec 01 20:45:01 crc kubenswrapper[4960]: I1201 20:45:01.782677 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" event={"ID":"fc634975-a76e-4aec-97e3-c1f1be39daa4","Type":"ContainerDied","Data":"4a5df75a0c4e3900b8b82218d6a256eb7122441fe2b3211e8ec31dc71cc490ac"} Dec 01 20:45:01 crc kubenswrapper[4960]: I1201 20:45:01.782728 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" event={"ID":"fc634975-a76e-4aec-97e3-c1f1be39daa4","Type":"ContainerStarted","Data":"1181c5707df93a0ff12082a2f76aaed90d3915106367b70ea239beb982076f5a"} Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.388296 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.493640 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume\") pod \"fc634975-a76e-4aec-97e3-c1f1be39daa4\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.493703 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zb82\" (UniqueName: \"kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82\") pod \"fc634975-a76e-4aec-97e3-c1f1be39daa4\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.493735 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume\") pod \"fc634975-a76e-4aec-97e3-c1f1be39daa4\" (UID: \"fc634975-a76e-4aec-97e3-c1f1be39daa4\") " Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.494495 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume" (OuterVolumeSpecName: "config-volume") pod "fc634975-a76e-4aec-97e3-c1f1be39daa4" (UID: "fc634975-a76e-4aec-97e3-c1f1be39daa4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.510782 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fc634975-a76e-4aec-97e3-c1f1be39daa4" (UID: "fc634975-a76e-4aec-97e3-c1f1be39daa4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.510955 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82" (OuterVolumeSpecName: "kube-api-access-8zb82") pod "fc634975-a76e-4aec-97e3-c1f1be39daa4" (UID: "fc634975-a76e-4aec-97e3-c1f1be39daa4"). InnerVolumeSpecName "kube-api-access-8zb82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.596033 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc634975-a76e-4aec-97e3-c1f1be39daa4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.596413 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zb82\" (UniqueName: \"kubernetes.io/projected/fc634975-a76e-4aec-97e3-c1f1be39daa4-kube-api-access-8zb82\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.596430 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc634975-a76e-4aec-97e3-c1f1be39daa4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.799867 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" event={"ID":"fc634975-a76e-4aec-97e3-c1f1be39daa4","Type":"ContainerDied","Data":"1181c5707df93a0ff12082a2f76aaed90d3915106367b70ea239beb982076f5a"} Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.799903 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1181c5707df93a0ff12082a2f76aaed90d3915106367b70ea239beb982076f5a" Dec 01 20:45:03 crc kubenswrapper[4960]: I1201 20:45:03.799950 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410365-8s8rj" Dec 01 20:45:04 crc kubenswrapper[4960]: I1201 20:45:04.470672 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg"] Dec 01 20:45:04 crc kubenswrapper[4960]: I1201 20:45:04.484939 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410320-lfrqg"] Dec 01 20:45:05 crc kubenswrapper[4960]: I1201 20:45:05.340279 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c107ba8b-ef9e-458f-a16f-f28f0625a85f" path="/var/lib/kubelet/pods/c107ba8b-ef9e-458f-a16f-f28f0625a85f/volumes" Dec 01 20:45:09 crc kubenswrapper[4960]: I1201 20:45:09.799404 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-dhxwf_ba9ca50d-24cd-4d42-b130-25c0e744bcb4/prometheus-operator/0.log" Dec 01 20:45:09 crc kubenswrapper[4960]: I1201 20:45:09.943316 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f/prometheus-operator-admission-webhook/0.log" Dec 01 20:45:10 crc kubenswrapper[4960]: I1201 20:45:10.078517 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_069dc362-519d-4349-b53e-288ee73b9916/prometheus-operator-admission-webhook/0.log" Dec 01 20:45:10 crc kubenswrapper[4960]: I1201 20:45:10.222167 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-gfmgc_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23/operator/0.log" Dec 01 20:45:10 crc kubenswrapper[4960]: I1201 20:45:10.329557 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tktnr_4f86337f-46f5-46aa-aaba-41179da85cab/perses-operator/0.log" Dec 01 20:45:12 crc kubenswrapper[4960]: I1201 20:45:12.324760 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:45:12 crc kubenswrapper[4960]: E1201 20:45:12.325475 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:45:16 crc kubenswrapper[4960]: I1201 20:45:16.501583 4960 scope.go:117] "RemoveContainer" containerID="cde6d9062ff642ba91612f15ab42c084f387dc3d584433060fa1ef0ea0896e53" Dec 01 20:45:24 crc kubenswrapper[4960]: I1201 20:45:24.327749 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:45:24 crc kubenswrapper[4960]: E1201 20:45:24.332586 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:45:25 crc kubenswrapper[4960]: I1201 20:45:25.479475 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/kube-rbac-proxy/0.log" Dec 01 20:45:25 crc kubenswrapper[4960]: I1201 20:45:25.524037 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/manager/0.log" Dec 01 20:45:37 crc kubenswrapper[4960]: I1201 20:45:37.325904 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:45:37 crc kubenswrapper[4960]: E1201 20:45:37.326656 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.210051 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:41 crc kubenswrapper[4960]: E1201 20:45:41.211173 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc634975-a76e-4aec-97e3-c1f1be39daa4" containerName="collect-profiles" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.211193 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc634975-a76e-4aec-97e3-c1f1be39daa4" containerName="collect-profiles" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.211470 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc634975-a76e-4aec-97e3-c1f1be39daa4" containerName="collect-profiles" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.215813 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.242086 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.405441 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.405558 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.405632 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zcjc\" (UniqueName: \"kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.506934 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.507023 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zcjc\" (UniqueName: \"kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.507162 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.507398 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.507647 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.530969 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zcjc\" (UniqueName: \"kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc\") pod \"redhat-marketplace-pgjsg\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:41 crc kubenswrapper[4960]: I1201 20:45:41.537924 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:42 crc kubenswrapper[4960]: I1201 20:45:42.100821 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:42 crc kubenswrapper[4960]: I1201 20:45:42.164846 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerStarted","Data":"36adc4fdd66b939aeb4e47516be57bf5345a343decf46cc51177a3e2314b4b22"} Dec 01 20:45:43 crc kubenswrapper[4960]: I1201 20:45:43.176008 4960 generic.go:334] "Generic (PLEG): container finished" podID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerID="4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209" exitCode=0 Dec 01 20:45:43 crc kubenswrapper[4960]: I1201 20:45:43.176171 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerDied","Data":"4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209"} Dec 01 20:45:43 crc kubenswrapper[4960]: I1201 20:45:43.181570 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:45:44 crc kubenswrapper[4960]: I1201 20:45:44.188577 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerStarted","Data":"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8"} Dec 01 20:45:45 crc kubenswrapper[4960]: I1201 20:45:45.200304 4960 generic.go:334] "Generic (PLEG): container finished" podID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerID="80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8" exitCode=0 Dec 01 20:45:45 crc kubenswrapper[4960]: I1201 20:45:45.200368 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerDied","Data":"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8"} Dec 01 20:45:46 crc kubenswrapper[4960]: I1201 20:45:46.213648 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerStarted","Data":"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9"} Dec 01 20:45:46 crc kubenswrapper[4960]: I1201 20:45:46.257250 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pgjsg" podStartSLOduration=2.83015173 podStartE2EDuration="5.257222897s" podCreationTimestamp="2025-12-01 20:45:41 +0000 UTC" firstStartedPulling="2025-12-01 20:45:43.181358384 +0000 UTC m=+3978.468850053" lastFinishedPulling="2025-12-01 20:45:45.608429531 +0000 UTC m=+3980.895921220" observedRunningTime="2025-12-01 20:45:46.250801536 +0000 UTC m=+3981.538293215" watchObservedRunningTime="2025-12-01 20:45:46.257222897 +0000 UTC m=+3981.544714566" Dec 01 20:45:49 crc kubenswrapper[4960]: I1201 20:45:49.324692 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:45:49 crc kubenswrapper[4960]: E1201 20:45:49.325436 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:45:51 crc kubenswrapper[4960]: I1201 20:45:51.545062 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:51 crc kubenswrapper[4960]: I1201 20:45:51.545740 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:51 crc kubenswrapper[4960]: I1201 20:45:51.641164 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:52 crc kubenswrapper[4960]: I1201 20:45:52.325847 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:53 crc kubenswrapper[4960]: I1201 20:45:53.806241 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:54 crc kubenswrapper[4960]: I1201 20:45:54.313874 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pgjsg" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="registry-server" containerID="cri-o://a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9" gracePeriod=2 Dec 01 20:45:54 crc kubenswrapper[4960]: I1201 20:45:54.979831 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.083054 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content\") pod \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.083235 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zcjc\" (UniqueName: \"kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc\") pod \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.083363 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities\") pod \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\" (UID: \"10c22ff4-3f8e-438f-b0db-8ff539e796c5\") " Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.084370 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities" (OuterVolumeSpecName: "utilities") pod "10c22ff4-3f8e-438f-b0db-8ff539e796c5" (UID: "10c22ff4-3f8e-438f-b0db-8ff539e796c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.103405 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10c22ff4-3f8e-438f-b0db-8ff539e796c5" (UID: "10c22ff4-3f8e-438f-b0db-8ff539e796c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.185527 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.185570 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10c22ff4-3f8e-438f-b0db-8ff539e796c5-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.222364 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:45:55 crc kubenswrapper[4960]: E1201 20:45:55.222933 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="extract-content" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.222955 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="extract-content" Dec 01 20:45:55 crc kubenswrapper[4960]: E1201 20:45:55.222971 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="extract-utilities" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.222980 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="extract-utilities" Dec 01 20:45:55 crc kubenswrapper[4960]: E1201 20:45:55.223040 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="registry-server" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.223049 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="registry-server" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.223327 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerName="registry-server" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.225406 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.232732 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.331676 4960 generic.go:334] "Generic (PLEG): container finished" podID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" containerID="a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9" exitCode=0 Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.333011 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pgjsg" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.340003 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerDied","Data":"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9"} Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.340202 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pgjsg" event={"ID":"10c22ff4-3f8e-438f-b0db-8ff539e796c5","Type":"ContainerDied","Data":"36adc4fdd66b939aeb4e47516be57bf5345a343decf46cc51177a3e2314b4b22"} Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.340285 4960 scope.go:117] "RemoveContainer" containerID="a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.359934 4960 scope.go:117] "RemoveContainer" containerID="80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.389743 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.390051 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.390181 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cj2v\" (UniqueName: \"kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.492338 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.492715 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.492812 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cj2v\" (UniqueName: \"kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.493033 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.493560 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.785396 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc" (OuterVolumeSpecName: "kube-api-access-4zcjc") pod "10c22ff4-3f8e-438f-b0db-8ff539e796c5" (UID: "10c22ff4-3f8e-438f-b0db-8ff539e796c5"). InnerVolumeSpecName "kube-api-access-4zcjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.788011 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cj2v\" (UniqueName: \"kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v\") pod \"certified-operators-mk425\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.799548 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zcjc\" (UniqueName: \"kubernetes.io/projected/10c22ff4-3f8e-438f-b0db-8ff539e796c5-kube-api-access-4zcjc\") on node \"crc\" DevicePath \"\"" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.803786 4960 scope.go:117] "RemoveContainer" containerID="4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209" Dec 01 20:45:55 crc kubenswrapper[4960]: I1201 20:45:55.873878 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.136651 4960 scope.go:117] "RemoveContainer" containerID="a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9" Dec 01 20:45:56 crc kubenswrapper[4960]: E1201 20:45:56.139033 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9\": container with ID starting with a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9 not found: ID does not exist" containerID="a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.139064 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9"} err="failed to get container status \"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9\": rpc error: code = NotFound desc = could not find container \"a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9\": container with ID starting with a67e3f3173adc25297bf8437a3a9f95256128e9bbfe1906fd4e4fbf0e32d4cd9 not found: ID does not exist" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.139082 4960 scope.go:117] "RemoveContainer" containerID="80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8" Dec 01 20:45:56 crc kubenswrapper[4960]: E1201 20:45:56.139296 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8\": container with ID starting with 80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8 not found: ID does not exist" containerID="80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.139317 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8"} err="failed to get container status \"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8\": rpc error: code = NotFound desc = could not find container \"80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8\": container with ID starting with 80dfdd6998829d990221cffb5997002e90b953d83bb6a001a80c4186888603c8 not found: ID does not exist" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.139329 4960 scope.go:117] "RemoveContainer" containerID="4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209" Dec 01 20:45:56 crc kubenswrapper[4960]: E1201 20:45:56.139575 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209\": container with ID starting with 4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209 not found: ID does not exist" containerID="4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.139594 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209"} err="failed to get container status \"4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209\": rpc error: code = NotFound desc = could not find container \"4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209\": container with ID starting with 4388a9ef36ad3be6786e3ff068c32d2f9092a3253d4f1c598170b8ff6516f209 not found: ID does not exist" Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.211546 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.219977 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pgjsg"] Dec 01 20:45:56 crc kubenswrapper[4960]: I1201 20:45:56.495846 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:45:57 crc kubenswrapper[4960]: I1201 20:45:57.333742 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c22ff4-3f8e-438f-b0db-8ff539e796c5" path="/var/lib/kubelet/pods/10c22ff4-3f8e-438f-b0db-8ff539e796c5/volumes" Dec 01 20:45:57 crc kubenswrapper[4960]: I1201 20:45:57.351199 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerID="cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01" exitCode=0 Dec 01 20:45:57 crc kubenswrapper[4960]: I1201 20:45:57.351236 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerDied","Data":"cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01"} Dec 01 20:45:57 crc kubenswrapper[4960]: I1201 20:45:57.351279 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerStarted","Data":"7c106f25b3aa583c47bcbd166cf5b8310ef9cc664221f7286e40c495828b296c"} Dec 01 20:45:58 crc kubenswrapper[4960]: I1201 20:45:58.363055 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerStarted","Data":"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397"} Dec 01 20:45:59 crc kubenswrapper[4960]: I1201 20:45:59.373442 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerID="8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397" exitCode=0 Dec 01 20:45:59 crc kubenswrapper[4960]: I1201 20:45:59.373542 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerDied","Data":"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397"} Dec 01 20:46:00 crc kubenswrapper[4960]: I1201 20:46:00.384635 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerStarted","Data":"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e"} Dec 01 20:46:00 crc kubenswrapper[4960]: I1201 20:46:00.404966 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mk425" podStartSLOduration=2.740873496 podStartE2EDuration="5.40494258s" podCreationTimestamp="2025-12-01 20:45:55 +0000 UTC" firstStartedPulling="2025-12-01 20:45:57.353165924 +0000 UTC m=+3992.640657603" lastFinishedPulling="2025-12-01 20:46:00.017234988 +0000 UTC m=+3995.304726687" observedRunningTime="2025-12-01 20:46:00.401652956 +0000 UTC m=+3995.689144625" watchObservedRunningTime="2025-12-01 20:46:00.40494258 +0000 UTC m=+3995.692434279" Dec 01 20:46:02 crc kubenswrapper[4960]: I1201 20:46:02.324947 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:46:02 crc kubenswrapper[4960]: E1201 20:46:02.326905 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:46:05 crc kubenswrapper[4960]: I1201 20:46:05.874134 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:05 crc kubenswrapper[4960]: I1201 20:46:05.874768 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:05 crc kubenswrapper[4960]: I1201 20:46:05.933108 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:06 crc kubenswrapper[4960]: I1201 20:46:06.505416 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:06 crc kubenswrapper[4960]: I1201 20:46:06.552320 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:46:08 crc kubenswrapper[4960]: I1201 20:46:08.462527 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mk425" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="registry-server" containerID="cri-o://f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e" gracePeriod=2 Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.151530 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.293967 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cj2v\" (UniqueName: \"kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v\") pod \"2ed05157-272e-48bf-bfac-f9c37e8f6879\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.294088 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content\") pod \"2ed05157-272e-48bf-bfac-f9c37e8f6879\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.294233 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities\") pod \"2ed05157-272e-48bf-bfac-f9c37e8f6879\" (UID: \"2ed05157-272e-48bf-bfac-f9c37e8f6879\") " Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.295074 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities" (OuterVolumeSpecName: "utilities") pod "2ed05157-272e-48bf-bfac-f9c37e8f6879" (UID: "2ed05157-272e-48bf-bfac-f9c37e8f6879"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.300797 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v" (OuterVolumeSpecName: "kube-api-access-6cj2v") pod "2ed05157-272e-48bf-bfac-f9c37e8f6879" (UID: "2ed05157-272e-48bf-bfac-f9c37e8f6879"). InnerVolumeSpecName "kube-api-access-6cj2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.349785 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ed05157-272e-48bf-bfac-f9c37e8f6879" (UID: "2ed05157-272e-48bf-bfac-f9c37e8f6879"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.396037 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cj2v\" (UniqueName: \"kubernetes.io/projected/2ed05157-272e-48bf-bfac-f9c37e8f6879-kube-api-access-6cj2v\") on node \"crc\" DevicePath \"\"" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.396064 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.396073 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ed05157-272e-48bf-bfac-f9c37e8f6879-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.472999 4960 generic.go:334] "Generic (PLEG): container finished" podID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerID="f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e" exitCode=0 Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.473042 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerDied","Data":"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e"} Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.473080 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk425" event={"ID":"2ed05157-272e-48bf-bfac-f9c37e8f6879","Type":"ContainerDied","Data":"7c106f25b3aa583c47bcbd166cf5b8310ef9cc664221f7286e40c495828b296c"} Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.473077 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk425" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.473094 4960 scope.go:117] "RemoveContainer" containerID="f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.500752 4960 scope.go:117] "RemoveContainer" containerID="8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.513357 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.522677 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mk425"] Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.532376 4960 scope.go:117] "RemoveContainer" containerID="cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.584602 4960 scope.go:117] "RemoveContainer" containerID="f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e" Dec 01 20:46:09 crc kubenswrapper[4960]: E1201 20:46:09.585016 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e\": container with ID starting with f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e not found: ID does not exist" containerID="f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.585076 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e"} err="failed to get container status \"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e\": rpc error: code = NotFound desc = could not find container \"f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e\": container with ID starting with f38ce05b22298df4c76aaeb6081b07ace5c69510fccf1a75237dfcb1d4ce798e not found: ID does not exist" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.585101 4960 scope.go:117] "RemoveContainer" containerID="8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397" Dec 01 20:46:09 crc kubenswrapper[4960]: E1201 20:46:09.585507 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397\": container with ID starting with 8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397 not found: ID does not exist" containerID="8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.585553 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397"} err="failed to get container status \"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397\": rpc error: code = NotFound desc = could not find container \"8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397\": container with ID starting with 8efdb6578ee29a0823b908e05e996530e750000fcc030973e81bde23ee13f397 not found: ID does not exist" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.585584 4960 scope.go:117] "RemoveContainer" containerID="cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01" Dec 01 20:46:09 crc kubenswrapper[4960]: E1201 20:46:09.585907 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01\": container with ID starting with cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01 not found: ID does not exist" containerID="cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01" Dec 01 20:46:09 crc kubenswrapper[4960]: I1201 20:46:09.585932 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01"} err="failed to get container status \"cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01\": rpc error: code = NotFound desc = could not find container \"cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01\": container with ID starting with cdb1e02b2c3a4ae0ffacc5454e8d7a9a0588cd8288e3bc0da687b44570416e01 not found: ID does not exist" Dec 01 20:46:11 crc kubenswrapper[4960]: I1201 20:46:11.338412 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" path="/var/lib/kubelet/pods/2ed05157-272e-48bf-bfac-f9c37e8f6879/volumes" Dec 01 20:46:13 crc kubenswrapper[4960]: I1201 20:46:13.323706 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:46:13 crc kubenswrapper[4960]: E1201 20:46:13.325323 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:46:26 crc kubenswrapper[4960]: I1201 20:46:26.324818 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:46:26 crc kubenswrapper[4960]: E1201 20:46:26.325928 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:46:37 crc kubenswrapper[4960]: I1201 20:46:37.339448 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:46:37 crc kubenswrapper[4960]: E1201 20:46:37.340271 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:46:50 crc kubenswrapper[4960]: I1201 20:46:50.324088 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:46:50 crc kubenswrapper[4960]: E1201 20:46:50.324879 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.147888 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:46:54 crc kubenswrapper[4960]: E1201 20:46:54.148891 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="extract-utilities" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.148905 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="extract-utilities" Dec 01 20:46:54 crc kubenswrapper[4960]: E1201 20:46:54.148913 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="extract-content" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.148919 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="extract-content" Dec 01 20:46:54 crc kubenswrapper[4960]: E1201 20:46:54.148934 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="registry-server" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.148941 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="registry-server" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.149138 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ed05157-272e-48bf-bfac-f9c37e8f6879" containerName="registry-server" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.150755 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.173551 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.309142 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.309261 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4n69\" (UniqueName: \"kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.309323 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.410749 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.411175 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4n69\" (UniqueName: \"kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.411229 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.411253 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.412022 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.436551 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4n69\" (UniqueName: \"kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69\") pod \"community-operators-vrp6j\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:54 crc kubenswrapper[4960]: I1201 20:46:54.515524 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:46:55 crc kubenswrapper[4960]: I1201 20:46:55.056724 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:46:55 crc kubenswrapper[4960]: I1201 20:46:55.954879 4960 generic.go:334] "Generic (PLEG): container finished" podID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerID="1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862" exitCode=0 Dec 01 20:46:55 crc kubenswrapper[4960]: I1201 20:46:55.955072 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerDied","Data":"1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862"} Dec 01 20:46:55 crc kubenswrapper[4960]: I1201 20:46:55.955200 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerStarted","Data":"387f1fd86fbded304eb2c8acb0deccd44bb0024fd83c2e6d4d6e22b278a505c1"} Dec 01 20:46:56 crc kubenswrapper[4960]: I1201 20:46:56.970244 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerStarted","Data":"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d"} Dec 01 20:46:57 crc kubenswrapper[4960]: I1201 20:46:57.986022 4960 generic.go:334] "Generic (PLEG): container finished" podID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerID="13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d" exitCode=0 Dec 01 20:46:57 crc kubenswrapper[4960]: I1201 20:46:57.986099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerDied","Data":"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d"} Dec 01 20:47:00 crc kubenswrapper[4960]: I1201 20:47:00.007868 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerStarted","Data":"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241"} Dec 01 20:47:00 crc kubenswrapper[4960]: I1201 20:47:00.027421 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vrp6j" podStartSLOduration=3.5706155649999998 podStartE2EDuration="6.027401937s" podCreationTimestamp="2025-12-01 20:46:54 +0000 UTC" firstStartedPulling="2025-12-01 20:46:55.956561712 +0000 UTC m=+4051.244053381" lastFinishedPulling="2025-12-01 20:46:58.413348084 +0000 UTC m=+4053.700839753" observedRunningTime="2025-12-01 20:47:00.022926605 +0000 UTC m=+4055.310418274" watchObservedRunningTime="2025-12-01 20:47:00.027401937 +0000 UTC m=+4055.314893606" Dec 01 20:47:03 crc kubenswrapper[4960]: I1201 20:47:03.324772 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:47:03 crc kubenswrapper[4960]: E1201 20:47:03.325343 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:47:04 crc kubenswrapper[4960]: I1201 20:47:04.517897 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:04 crc kubenswrapper[4960]: I1201 20:47:04.518515 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:04 crc kubenswrapper[4960]: I1201 20:47:04.566468 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:04 crc kubenswrapper[4960]: I1201 20:47:04.831401 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:04 crc kubenswrapper[4960]: I1201 20:47:04.886780 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:47:06 crc kubenswrapper[4960]: I1201 20:47:06.790222 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vrp6j" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="registry-server" containerID="cri-o://f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241" gracePeriod=2 Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.491071 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.531535 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities\") pod \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.531608 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4n69\" (UniqueName: \"kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69\") pod \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.531708 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content\") pod \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\" (UID: \"0c511078-f423-4ebe-af2c-bd80a4c4c82f\") " Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.533030 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities" (OuterVolumeSpecName: "utilities") pod "0c511078-f423-4ebe-af2c-bd80a4c4c82f" (UID: "0c511078-f423-4ebe-af2c-bd80a4c4c82f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.533659 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.538101 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69" (OuterVolumeSpecName: "kube-api-access-c4n69") pod "0c511078-f423-4ebe-af2c-bd80a4c4c82f" (UID: "0c511078-f423-4ebe-af2c-bd80a4c4c82f"). InnerVolumeSpecName "kube-api-access-c4n69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.582710 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c511078-f423-4ebe-af2c-bd80a4c4c82f" (UID: "0c511078-f423-4ebe-af2c-bd80a4c4c82f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.635499 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c511078-f423-4ebe-af2c-bd80a4c4c82f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.635535 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4n69\" (UniqueName: \"kubernetes.io/projected/0c511078-f423-4ebe-af2c-bd80a4c4c82f-kube-api-access-c4n69\") on node \"crc\" DevicePath \"\"" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.800784 4960 generic.go:334] "Generic (PLEG): container finished" podID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerID="f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241" exitCode=0 Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.800824 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerDied","Data":"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241"} Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.800848 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vrp6j" event={"ID":"0c511078-f423-4ebe-af2c-bd80a4c4c82f","Type":"ContainerDied","Data":"387f1fd86fbded304eb2c8acb0deccd44bb0024fd83c2e6d4d6e22b278a505c1"} Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.800864 4960 scope.go:117] "RemoveContainer" containerID="f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.800974 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vrp6j" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.846401 4960 scope.go:117] "RemoveContainer" containerID="13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.850264 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.869763 4960 scope.go:117] "RemoveContainer" containerID="1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.887460 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vrp6j"] Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.921069 4960 scope.go:117] "RemoveContainer" containerID="f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241" Dec 01 20:47:07 crc kubenswrapper[4960]: E1201 20:47:07.921907 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241\": container with ID starting with f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241 not found: ID does not exist" containerID="f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.921940 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241"} err="failed to get container status \"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241\": rpc error: code = NotFound desc = could not find container \"f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241\": container with ID starting with f3fb7fadba83775cfc38cdbbe7730695b9e92cdec6d4bb031b14ccb7dd0de241 not found: ID does not exist" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.921964 4960 scope.go:117] "RemoveContainer" containerID="13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d" Dec 01 20:47:07 crc kubenswrapper[4960]: E1201 20:47:07.922518 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d\": container with ID starting with 13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d not found: ID does not exist" containerID="13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.922546 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d"} err="failed to get container status \"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d\": rpc error: code = NotFound desc = could not find container \"13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d\": container with ID starting with 13b180fe202444c96c5ab74592460508166ebfe63c3ae9ffb4d93bbe297a3f5d not found: ID does not exist" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.922560 4960 scope.go:117] "RemoveContainer" containerID="1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862" Dec 01 20:47:07 crc kubenswrapper[4960]: E1201 20:47:07.923049 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862\": container with ID starting with 1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862 not found: ID does not exist" containerID="1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862" Dec 01 20:47:07 crc kubenswrapper[4960]: I1201 20:47:07.923071 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862"} err="failed to get container status \"1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862\": rpc error: code = NotFound desc = could not find container \"1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862\": container with ID starting with 1354ed9ce174eb1f9fa90556492654840b66d2e6862f7133bcfbaeea84da0862 not found: ID does not exist" Dec 01 20:47:09 crc kubenswrapper[4960]: I1201 20:47:09.346001 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" path="/var/lib/kubelet/pods/0c511078-f423-4ebe-af2c-bd80a4c4c82f/volumes" Dec 01 20:47:14 crc kubenswrapper[4960]: I1201 20:47:14.324149 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:47:14 crc kubenswrapper[4960]: E1201 20:47:14.324875 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:47:15 crc kubenswrapper[4960]: I1201 20:47:15.889424 4960 generic.go:334] "Generic (PLEG): container finished" podID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerID="d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e" exitCode=0 Dec 01 20:47:15 crc kubenswrapper[4960]: I1201 20:47:15.889525 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-khtrz/must-gather-zq59t" event={"ID":"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e","Type":"ContainerDied","Data":"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e"} Dec 01 20:47:15 crc kubenswrapper[4960]: I1201 20:47:15.890299 4960 scope.go:117] "RemoveContainer" containerID="d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e" Dec 01 20:47:16 crc kubenswrapper[4960]: I1201 20:47:16.172537 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-khtrz_must-gather-zq59t_0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e/gather/0.log" Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.032722 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-khtrz/must-gather-zq59t"] Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.033565 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-khtrz/must-gather-zq59t" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="copy" containerID="cri-o://372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0" gracePeriod=2 Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.042608 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-khtrz/must-gather-zq59t"] Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.857532 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-khtrz_must-gather-zq59t_0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e/copy/0.log" Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.859161 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.985833 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-khtrz_must-gather-zq59t_0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e/copy/0.log" Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.986149 4960 generic.go:334] "Generic (PLEG): container finished" podID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerID="372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0" exitCode=143 Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.986204 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-khtrz/must-gather-zq59t" Dec 01 20:47:25 crc kubenswrapper[4960]: I1201 20:47:25.986210 4960 scope.go:117] "RemoveContainer" containerID="372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.005543 4960 scope.go:117] "RemoveContainer" containerID="d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.051345 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output\") pod \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.051440 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68tfr\" (UniqueName: \"kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr\") pod \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\" (UID: \"0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e\") " Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.052407 4960 scope.go:117] "RemoveContainer" containerID="372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0" Dec 01 20:47:26 crc kubenswrapper[4960]: E1201 20:47:26.053762 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0\": container with ID starting with 372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0 not found: ID does not exist" containerID="372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.053794 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0"} err="failed to get container status \"372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0\": rpc error: code = NotFound desc = could not find container \"372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0\": container with ID starting with 372637eb2f34a0c5d745919511b0089b5be782ccec5866c7ebae7ca2ce9a4fa0 not found: ID does not exist" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.053815 4960 scope.go:117] "RemoveContainer" containerID="d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e" Dec 01 20:47:26 crc kubenswrapper[4960]: E1201 20:47:26.054344 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e\": container with ID starting with d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e not found: ID does not exist" containerID="d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.054393 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e"} err="failed to get container status \"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e\": rpc error: code = NotFound desc = could not find container \"d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e\": container with ID starting with d3ef81ea3e674c722916400afd58ddb7baba5cc5fbf741a008286c0bd2b8de1e not found: ID does not exist" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.057387 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr" (OuterVolumeSpecName: "kube-api-access-68tfr") pod "0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" (UID: "0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e"). InnerVolumeSpecName "kube-api-access-68tfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.153914 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68tfr\" (UniqueName: \"kubernetes.io/projected/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-kube-api-access-68tfr\") on node \"crc\" DevicePath \"\"" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.223915 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" (UID: "0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:47:26 crc kubenswrapper[4960]: I1201 20:47:26.255736 4960 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 20:47:27 crc kubenswrapper[4960]: I1201 20:47:27.336465 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" path="/var/lib/kubelet/pods/0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e/volumes" Dec 01 20:47:28 crc kubenswrapper[4960]: I1201 20:47:28.323989 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:47:28 crc kubenswrapper[4960]: E1201 20:47:28.324288 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:47:40 crc kubenswrapper[4960]: I1201 20:47:40.323973 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:47:40 crc kubenswrapper[4960]: E1201 20:47:40.324939 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.146630 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:47:46 crc kubenswrapper[4960]: E1201 20:47:46.147600 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="extract-utilities" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147618 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="extract-utilities" Dec 01 20:47:46 crc kubenswrapper[4960]: E1201 20:47:46.147638 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="copy" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147646 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="copy" Dec 01 20:47:46 crc kubenswrapper[4960]: E1201 20:47:46.147656 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="registry-server" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147664 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="registry-server" Dec 01 20:47:46 crc kubenswrapper[4960]: E1201 20:47:46.147698 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="gather" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147704 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="gather" Dec 01 20:47:46 crc kubenswrapper[4960]: E1201 20:47:46.147725 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="extract-content" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147734 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="extract-content" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147964 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="copy" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147982 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c511078-f423-4ebe-af2c-bd80a4c4c82f" containerName="registry-server" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.147999 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e7c0c3e-c8a7-4060-9ea0-79eff930dd5e" containerName="gather" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.149686 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.165687 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.267352 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.267439 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6wwz\" (UniqueName: \"kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.267512 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.370191 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.370277 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6wwz\" (UniqueName: \"kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.370341 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.370712 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.370723 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.391783 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6wwz\" (UniqueName: \"kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz\") pod \"redhat-operators-67bql\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.473481 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:47:46 crc kubenswrapper[4960]: I1201 20:47:46.988477 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:47:47 crc kubenswrapper[4960]: I1201 20:47:47.176852 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerStarted","Data":"f9cfe1623c4cf1f169c1fc123a2c9ac437c1dc2262807c27d71ca3c4fb2d412e"} Dec 01 20:47:48 crc kubenswrapper[4960]: I1201 20:47:48.191770 4960 generic.go:334] "Generic (PLEG): container finished" podID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerID="7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe" exitCode=0 Dec 01 20:47:48 crc kubenswrapper[4960]: I1201 20:47:48.192099 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerDied","Data":"7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe"} Dec 01 20:47:53 crc kubenswrapper[4960]: I1201 20:47:53.325125 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:47:53 crc kubenswrapper[4960]: E1201 20:47:53.325765 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:47:55 crc kubenswrapper[4960]: I1201 20:47:55.265199 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerStarted","Data":"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a"} Dec 01 20:47:57 crc kubenswrapper[4960]: I1201 20:47:57.288236 4960 generic.go:334] "Generic (PLEG): container finished" podID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerID="6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a" exitCode=0 Dec 01 20:47:57 crc kubenswrapper[4960]: I1201 20:47:57.288334 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerDied","Data":"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a"} Dec 01 20:47:58 crc kubenswrapper[4960]: I1201 20:47:58.299480 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerStarted","Data":"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba"} Dec 01 20:47:58 crc kubenswrapper[4960]: I1201 20:47:58.327295 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-67bql" podStartSLOduration=2.565299825 podStartE2EDuration="12.327268163s" podCreationTimestamp="2025-12-01 20:47:46 +0000 UTC" firstStartedPulling="2025-12-01 20:47:48.193552407 +0000 UTC m=+4103.481044076" lastFinishedPulling="2025-12-01 20:47:57.955520745 +0000 UTC m=+4113.243012414" observedRunningTime="2025-12-01 20:47:58.316541205 +0000 UTC m=+4113.604032874" watchObservedRunningTime="2025-12-01 20:47:58.327268163 +0000 UTC m=+4113.614759832" Dec 01 20:48:06 crc kubenswrapper[4960]: I1201 20:48:06.473775 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:06 crc kubenswrapper[4960]: I1201 20:48:06.474389 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:07 crc kubenswrapper[4960]: I1201 20:48:07.230210 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:07 crc kubenswrapper[4960]: I1201 20:48:07.452015 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:07 crc kubenswrapper[4960]: I1201 20:48:07.508469 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:48:08 crc kubenswrapper[4960]: I1201 20:48:08.324128 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:48:08 crc kubenswrapper[4960]: E1201 20:48:08.324409 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:48:09 crc kubenswrapper[4960]: I1201 20:48:09.416516 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-67bql" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="registry-server" containerID="cri-o://825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba" gracePeriod=2 Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.007342 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.104978 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content\") pod \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.105277 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities\") pod \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.105341 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6wwz\" (UniqueName: \"kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz\") pod \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\" (UID: \"0ce2a81e-aee2-4f8b-8652-9941541a2a49\") " Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.106262 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities" (OuterVolumeSpecName: "utilities") pod "0ce2a81e-aee2-4f8b-8652-9941541a2a49" (UID: "0ce2a81e-aee2-4f8b-8652-9941541a2a49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.112455 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz" (OuterVolumeSpecName: "kube-api-access-t6wwz") pod "0ce2a81e-aee2-4f8b-8652-9941541a2a49" (UID: "0ce2a81e-aee2-4f8b-8652-9941541a2a49"). InnerVolumeSpecName "kube-api-access-t6wwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.207925 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.207977 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6wwz\" (UniqueName: \"kubernetes.io/projected/0ce2a81e-aee2-4f8b-8652-9941541a2a49-kube-api-access-t6wwz\") on node \"crc\" DevicePath \"\"" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.210566 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ce2a81e-aee2-4f8b-8652-9941541a2a49" (UID: "0ce2a81e-aee2-4f8b-8652-9941541a2a49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.310069 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce2a81e-aee2-4f8b-8652-9941541a2a49-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.426642 4960 generic.go:334] "Generic (PLEG): container finished" podID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerID="825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba" exitCode=0 Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.426688 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerDied","Data":"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba"} Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.426949 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-67bql" event={"ID":"0ce2a81e-aee2-4f8b-8652-9941541a2a49","Type":"ContainerDied","Data":"f9cfe1623c4cf1f169c1fc123a2c9ac437c1dc2262807c27d71ca3c4fb2d412e"} Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.426976 4960 scope.go:117] "RemoveContainer" containerID="825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.426733 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-67bql" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.459767 4960 scope.go:117] "RemoveContainer" containerID="6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.468995 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.480665 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-67bql"] Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.492851 4960 scope.go:117] "RemoveContainer" containerID="7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.526553 4960 scope.go:117] "RemoveContainer" containerID="825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba" Dec 01 20:48:10 crc kubenswrapper[4960]: E1201 20:48:10.527063 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba\": container with ID starting with 825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba not found: ID does not exist" containerID="825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.527165 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba"} err="failed to get container status \"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba\": rpc error: code = NotFound desc = could not find container \"825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba\": container with ID starting with 825d94fcd22c0ee92c2528d216c45ddb05701d706d05f11e31418f8c1b35ddba not found: ID does not exist" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.527200 4960 scope.go:117] "RemoveContainer" containerID="6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a" Dec 01 20:48:10 crc kubenswrapper[4960]: E1201 20:48:10.527671 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a\": container with ID starting with 6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a not found: ID does not exist" containerID="6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.527714 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a"} err="failed to get container status \"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a\": rpc error: code = NotFound desc = could not find container \"6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a\": container with ID starting with 6375ab1c5c98a4165974c2d9ae6b90e900d893bbf0409f02c6b84acfe98e0e1a not found: ID does not exist" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.527742 4960 scope.go:117] "RemoveContainer" containerID="7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe" Dec 01 20:48:10 crc kubenswrapper[4960]: E1201 20:48:10.527985 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe\": container with ID starting with 7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe not found: ID does not exist" containerID="7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe" Dec 01 20:48:10 crc kubenswrapper[4960]: I1201 20:48:10.528020 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe"} err="failed to get container status \"7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe\": rpc error: code = NotFound desc = could not find container \"7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe\": container with ID starting with 7b0af2733c018ce3ea29f43db0c0f2537f7f91d2ae5427fdcbf0d96f9eed5abe not found: ID does not exist" Dec 01 20:48:11 crc kubenswrapper[4960]: I1201 20:48:11.340522 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" path="/var/lib/kubelet/pods/0ce2a81e-aee2-4f8b-8652-9941541a2a49/volumes" Dec 01 20:48:22 crc kubenswrapper[4960]: I1201 20:48:22.324683 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:48:22 crc kubenswrapper[4960]: E1201 20:48:22.325496 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:48:35 crc kubenswrapper[4960]: I1201 20:48:35.336451 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:48:35 crc kubenswrapper[4960]: E1201 20:48:35.338778 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:48:47 crc kubenswrapper[4960]: I1201 20:48:47.325610 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:48:47 crc kubenswrapper[4960]: I1201 20:48:47.839167 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786"} Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.614328 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ztv6/must-gather-pjwp5"] Dec 01 20:50:18 crc kubenswrapper[4960]: E1201 20:50:18.615438 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="extract-utilities" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.615459 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="extract-utilities" Dec 01 20:50:18 crc kubenswrapper[4960]: E1201 20:50:18.615479 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="registry-server" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.615488 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="registry-server" Dec 01 20:50:18 crc kubenswrapper[4960]: E1201 20:50:18.615506 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="extract-content" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.615514 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="extract-content" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.615758 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce2a81e-aee2-4f8b-8652-9941541a2a49" containerName="registry-server" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.622775 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.625513 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ztv6"/"kube-root-ca.crt" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.625728 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ztv6"/"openshift-service-ca.crt" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.644657 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ztv6/must-gather-pjwp5"] Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.732267 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvj27\" (UniqueName: \"kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.732357 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.834260 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvj27\" (UniqueName: \"kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.834380 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.834845 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.871028 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvj27\" (UniqueName: \"kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27\") pod \"must-gather-pjwp5\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:18 crc kubenswrapper[4960]: I1201 20:50:18.945393 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:50:19 crc kubenswrapper[4960]: I1201 20:50:19.441006 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ztv6/must-gather-pjwp5"] Dec 01 20:50:19 crc kubenswrapper[4960]: I1201 20:50:19.721242 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" event={"ID":"31fa442c-a776-4f31-9a4b-44ac89b0ff31","Type":"ContainerStarted","Data":"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c"} Dec 01 20:50:19 crc kubenswrapper[4960]: I1201 20:50:19.721567 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" event={"ID":"31fa442c-a776-4f31-9a4b-44ac89b0ff31","Type":"ContainerStarted","Data":"56bf3a6cfbab1b20f40ec97cfdb817115c9be3823a75bd01fe4477f4943795ee"} Dec 01 20:50:20 crc kubenswrapper[4960]: I1201 20:50:20.730539 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" event={"ID":"31fa442c-a776-4f31-9a4b-44ac89b0ff31","Type":"ContainerStarted","Data":"2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6"} Dec 01 20:50:20 crc kubenswrapper[4960]: I1201 20:50:20.754627 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" podStartSLOduration=2.7546106999999997 podStartE2EDuration="2.7546107s" podCreationTimestamp="2025-12-01 20:50:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-01 20:50:20.74762196 +0000 UTC m=+4256.035113629" watchObservedRunningTime="2025-12-01 20:50:20.7546107 +0000 UTC m=+4256.042102369" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.540397 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-rtbs4"] Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.544992 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.546720 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8ztv6"/"default-dockercfg-88khk" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.632916 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.632967 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl8gz\" (UniqueName: \"kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.735375 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.735447 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl8gz\" (UniqueName: \"kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.736051 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.758904 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl8gz\" (UniqueName: \"kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz\") pod \"crc-debug-rtbs4\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: I1201 20:50:23.863126 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:50:23 crc kubenswrapper[4960]: W1201 20:50:23.897687 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46cc94e9_e28b_46ee_92b7_93c8727b15fe.slice/crio-b9b3aa83a94b8bf6f09e68203dd3c2a8decf4af570b8efa2792e75e92db8f5c3 WatchSource:0}: Error finding container b9b3aa83a94b8bf6f09e68203dd3c2a8decf4af570b8efa2792e75e92db8f5c3: Status 404 returned error can't find the container with id b9b3aa83a94b8bf6f09e68203dd3c2a8decf4af570b8efa2792e75e92db8f5c3 Dec 01 20:50:24 crc kubenswrapper[4960]: I1201 20:50:24.795693 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" event={"ID":"46cc94e9-e28b-46ee-92b7-93c8727b15fe","Type":"ContainerStarted","Data":"8b3ad2b57d97604a86c0af119eaabba75e7a50f3b9d24584baaba18c0c3c7c41"} Dec 01 20:50:24 crc kubenswrapper[4960]: I1201 20:50:24.796217 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" event={"ID":"46cc94e9-e28b-46ee-92b7-93c8727b15fe","Type":"ContainerStarted","Data":"b9b3aa83a94b8bf6f09e68203dd3c2a8decf4af570b8efa2792e75e92db8f5c3"} Dec 01 20:51:05 crc kubenswrapper[4960]: I1201 20:51:05.204731 4960 generic.go:334] "Generic (PLEG): container finished" podID="46cc94e9-e28b-46ee-92b7-93c8727b15fe" containerID="8b3ad2b57d97604a86c0af119eaabba75e7a50f3b9d24584baaba18c0c3c7c41" exitCode=0 Dec 01 20:51:05 crc kubenswrapper[4960]: I1201 20:51:05.204895 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" event={"ID":"46cc94e9-e28b-46ee-92b7-93c8727b15fe","Type":"ContainerDied","Data":"8b3ad2b57d97604a86c0af119eaabba75e7a50f3b9d24584baaba18c0c3c7c41"} Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.714691 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.750441 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-rtbs4"] Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.762077 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-rtbs4"] Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.768332 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host\") pod \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.768445 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host" (OuterVolumeSpecName: "host") pod "46cc94e9-e28b-46ee-92b7-93c8727b15fe" (UID: "46cc94e9-e28b-46ee-92b7-93c8727b15fe"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.768718 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl8gz\" (UniqueName: \"kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz\") pod \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\" (UID: \"46cc94e9-e28b-46ee-92b7-93c8727b15fe\") " Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.769330 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46cc94e9-e28b-46ee-92b7-93c8727b15fe-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.776348 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz" (OuterVolumeSpecName: "kube-api-access-jl8gz") pod "46cc94e9-e28b-46ee-92b7-93c8727b15fe" (UID: "46cc94e9-e28b-46ee-92b7-93c8727b15fe"). InnerVolumeSpecName "kube-api-access-jl8gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:51:06 crc kubenswrapper[4960]: I1201 20:51:06.872033 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl8gz\" (UniqueName: \"kubernetes.io/projected/46cc94e9-e28b-46ee-92b7-93c8727b15fe-kube-api-access-jl8gz\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.231833 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9b3aa83a94b8bf6f09e68203dd3c2a8decf4af570b8efa2792e75e92db8f5c3" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.232090 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-rtbs4" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.336099 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46cc94e9-e28b-46ee-92b7-93c8727b15fe" path="/var/lib/kubelet/pods/46cc94e9-e28b-46ee-92b7-93c8727b15fe/volumes" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.952704 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-2fg65"] Dec 01 20:51:07 crc kubenswrapper[4960]: E1201 20:51:07.953087 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46cc94e9-e28b-46ee-92b7-93c8727b15fe" containerName="container-00" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.953098 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="46cc94e9-e28b-46ee-92b7-93c8727b15fe" containerName="container-00" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.953373 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="46cc94e9-e28b-46ee-92b7-93c8727b15fe" containerName="container-00" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.954060 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.956439 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8ztv6"/"default-dockercfg-88khk" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.993488 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:07 crc kubenswrapper[4960]: I1201 20:51:07.993896 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhfsj\" (UniqueName: \"kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:08 crc kubenswrapper[4960]: I1201 20:51:08.096295 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhfsj\" (UniqueName: \"kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:08 crc kubenswrapper[4960]: I1201 20:51:08.096448 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:08 crc kubenswrapper[4960]: I1201 20:51:08.096645 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:08 crc kubenswrapper[4960]: I1201 20:51:08.493091 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhfsj\" (UniqueName: \"kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj\") pod \"crc-debug-2fg65\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:08 crc kubenswrapper[4960]: I1201 20:51:08.574438 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:09 crc kubenswrapper[4960]: I1201 20:51:09.251721 4960 generic.go:334] "Generic (PLEG): container finished" podID="5e25459d-e700-4711-905d-e9b5c1c2afed" containerID="f15ea7ebc9caa9551247852151f9d28cf3dce6ad5a15f89da6fc34c4b19992ff" exitCode=0 Dec 01 20:51:09 crc kubenswrapper[4960]: I1201 20:51:09.251827 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" event={"ID":"5e25459d-e700-4711-905d-e9b5c1c2afed","Type":"ContainerDied","Data":"f15ea7ebc9caa9551247852151f9d28cf3dce6ad5a15f89da6fc34c4b19992ff"} Dec 01 20:51:09 crc kubenswrapper[4960]: I1201 20:51:09.252246 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" event={"ID":"5e25459d-e700-4711-905d-e9b5c1c2afed","Type":"ContainerStarted","Data":"b6d2d30cfb70e471a6836fcdb0bec27cf427c0bce373c9b5cadf00294ce4f736"} Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.388318 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.388724 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-2fg65"] Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.403243 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-2fg65"] Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.452302 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhfsj\" (UniqueName: \"kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj\") pod \"5e25459d-e700-4711-905d-e9b5c1c2afed\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.452589 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host\") pod \"5e25459d-e700-4711-905d-e9b5c1c2afed\" (UID: \"5e25459d-e700-4711-905d-e9b5c1c2afed\") " Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.453624 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host" (OuterVolumeSpecName: "host") pod "5e25459d-e700-4711-905d-e9b5c1c2afed" (UID: "5e25459d-e700-4711-905d-e9b5c1c2afed"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.461964 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj" (OuterVolumeSpecName: "kube-api-access-vhfsj") pod "5e25459d-e700-4711-905d-e9b5c1c2afed" (UID: "5e25459d-e700-4711-905d-e9b5c1c2afed"). InnerVolumeSpecName "kube-api-access-vhfsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.554935 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e25459d-e700-4711-905d-e9b5c1c2afed-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:10 crc kubenswrapper[4960]: I1201 20:51:10.555129 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhfsj\" (UniqueName: \"kubernetes.io/projected/5e25459d-e700-4711-905d-e9b5c1c2afed-kube-api-access-vhfsj\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.275944 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6d2d30cfb70e471a6836fcdb0bec27cf427c0bce373c9b5cadf00294ce4f736" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.275995 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-2fg65" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.336575 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e25459d-e700-4711-905d-e9b5c1c2afed" path="/var/lib/kubelet/pods/5e25459d-e700-4711-905d-e9b5c1c2afed/volumes" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.592233 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-xp42k"] Dec 01 20:51:11 crc kubenswrapper[4960]: E1201 20:51:11.592678 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e25459d-e700-4711-905d-e9b5c1c2afed" containerName="container-00" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.592689 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e25459d-e700-4711-905d-e9b5c1c2afed" containerName="container-00" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.592893 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e25459d-e700-4711-905d-e9b5c1c2afed" containerName="container-00" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.593624 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.596237 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8ztv6"/"default-dockercfg-88khk" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.678257 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.678334 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2bzf\" (UniqueName: \"kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.781143 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.781226 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2bzf\" (UniqueName: \"kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.781296 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.809918 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2bzf\" (UniqueName: \"kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf\") pod \"crc-debug-xp42k\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: I1201 20:51:11.912618 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:11 crc kubenswrapper[4960]: W1201 20:51:11.940875 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d3cb255_1f7e_4256_bbf3_80b208a265db.slice/crio-b3d82a3cec916cba624bd661f6ee6506708226316833a700d88472990223a807 WatchSource:0}: Error finding container b3d82a3cec916cba624bd661f6ee6506708226316833a700d88472990223a807: Status 404 returned error can't find the container with id b3d82a3cec916cba624bd661f6ee6506708226316833a700d88472990223a807 Dec 01 20:51:12 crc kubenswrapper[4960]: I1201 20:51:12.285304 4960 generic.go:334] "Generic (PLEG): container finished" podID="9d3cb255-1f7e-4256-bbf3-80b208a265db" containerID="6eaa9ecc1ec67e16ad00e28218ba9e1da1bd2bc02ce3cd980247f31880f5106c" exitCode=0 Dec 01 20:51:12 crc kubenswrapper[4960]: I1201 20:51:12.285489 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" event={"ID":"9d3cb255-1f7e-4256-bbf3-80b208a265db","Type":"ContainerDied","Data":"6eaa9ecc1ec67e16ad00e28218ba9e1da1bd2bc02ce3cd980247f31880f5106c"} Dec 01 20:51:12 crc kubenswrapper[4960]: I1201 20:51:12.285652 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" event={"ID":"9d3cb255-1f7e-4256-bbf3-80b208a265db","Type":"ContainerStarted","Data":"b3d82a3cec916cba624bd661f6ee6506708226316833a700d88472990223a807"} Dec 01 20:51:12 crc kubenswrapper[4960]: I1201 20:51:12.335997 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-xp42k"] Dec 01 20:51:12 crc kubenswrapper[4960]: I1201 20:51:12.349105 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ztv6/crc-debug-xp42k"] Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.437978 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.532734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host\") pod \"9d3cb255-1f7e-4256-bbf3-80b208a265db\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.533017 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2bzf\" (UniqueName: \"kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf\") pod \"9d3cb255-1f7e-4256-bbf3-80b208a265db\" (UID: \"9d3cb255-1f7e-4256-bbf3-80b208a265db\") " Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.534545 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host" (OuterVolumeSpecName: "host") pod "9d3cb255-1f7e-4256-bbf3-80b208a265db" (UID: "9d3cb255-1f7e-4256-bbf3-80b208a265db"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.539951 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf" (OuterVolumeSpecName: "kube-api-access-q2bzf") pod "9d3cb255-1f7e-4256-bbf3-80b208a265db" (UID: "9d3cb255-1f7e-4256-bbf3-80b208a265db"). InnerVolumeSpecName "kube-api-access-q2bzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.636065 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2bzf\" (UniqueName: \"kubernetes.io/projected/9d3cb255-1f7e-4256-bbf3-80b208a265db-kube-api-access-q2bzf\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.636134 4960 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9d3cb255-1f7e-4256-bbf3-80b208a265db-host\") on node \"crc\" DevicePath \"\"" Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.870941 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:51:13 crc kubenswrapper[4960]: I1201 20:51:13.871337 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:51:14 crc kubenswrapper[4960]: I1201 20:51:14.328652 4960 scope.go:117] "RemoveContainer" containerID="6eaa9ecc1ec67e16ad00e28218ba9e1da1bd2bc02ce3cd980247f31880f5106c" Dec 01 20:51:14 crc kubenswrapper[4960]: I1201 20:51:14.328679 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/crc-debug-xp42k" Dec 01 20:51:15 crc kubenswrapper[4960]: I1201 20:51:15.346418 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3cb255-1f7e-4256-bbf3-80b208a265db" path="/var/lib/kubelet/pods/9d3cb255-1f7e-4256-bbf3-80b208a265db/volumes" Dec 01 20:51:43 crc kubenswrapper[4960]: I1201 20:51:43.871227 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:51:43 crc kubenswrapper[4960]: I1201 20:51:43.872798 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:51:50 crc kubenswrapper[4960]: I1201 20:51:50.998642 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/init-config-reloader/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.150830 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/alertmanager/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.182062 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/init-config-reloader/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.193605 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_39fbd706-99ef-4b5d-9d7a-e84ee337c443/config-reloader/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.554805 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d89589d9d-9dglh_9c816e6d-aa4e-46cc-8494-39d3f334a028/barbican-api/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.571734 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d89589d9d-9dglh_9c816e6d-aa4e-46cc-8494-39d3f334a028/barbican-api-log/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.659520 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f67669b46-khzh6_4771a4c0-c72c-4819-953e-ed6ecd9d1ef6/barbican-keystone-listener/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.871590 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f67669b46-khzh6_4771a4c0-c72c-4819-953e-ed6ecd9d1ef6/barbican-keystone-listener-log/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.874987 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f647bc89f-mwqdp_c6e92e0f-eb53-4d9a-af6a-4294f774a870/barbican-worker/0.log" Dec 01 20:51:51 crc kubenswrapper[4960]: I1201 20:51:51.930779 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f647bc89f-mwqdp_c6e92e0f-eb53-4d9a-af6a-4294f774a870/barbican-worker-log/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.139924 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-gs9jb_098a01da-e6f9-4bc8-8249-72bbeaf949f7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.215623 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/ceilometer-central-agent/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.371317 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/ceilometer-notification-agent/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.386894 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/sg-core/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.418514 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_62742f95-55dd-471c-964f-33b6eea63418/proxy-httpd/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.599439 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b45703c2-5b71-47a4-835a-cee374c228b7/cinder-api/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.618830 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b45703c2-5b71-47a4-835a-cee374c228b7/cinder-api-log/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.743033 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c0bee6fd-952d-4a33-83dc-c93625f15979/cinder-scheduler/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.848190 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c0bee6fd-952d-4a33-83dc-c93625f15979/probe/0.log" Dec 01 20:51:52 crc kubenswrapper[4960]: I1201 20:51:52.983633 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf/cloudkitty-api-log/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.008677 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-api-0_9e26ecfd-4929-4047-a4a8-6ce5b5d9a3bf/cloudkitty-api/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.139824 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-compactor-0_8e8a9388-578d-4d41-a2cf-d2ad6e344288/loki-compactor/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.234306 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-distributor-56cd74f89f-jxqfj_2e566703-0d27-4306-bd42-7a91da29c736/loki-distributor/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.353349 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-mpddm_5cc4eeaa-88b7-4439-a39f-5e889845181e/gateway/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.517192 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-gateway-76cc998948-xw4bs_a8b0ac50-f375-427e-b103-0bfd272201ea/gateway/0.log" Dec 01 20:51:53 crc kubenswrapper[4960]: I1201 20:51:53.672302 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-index-gateway-0_eaed0cb9-fd50-4f3a-9ca1-8c99ecbafc38/loki-index-gateway/0.log" Dec 01 20:51:54 crc kubenswrapper[4960]: I1201 20:51:54.081603 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-ingester-0_969241f0-8883-4570-94d4-fdb20ee094cd/loki-ingester/0.log" Dec 01 20:51:54 crc kubenswrapper[4960]: I1201 20:51:54.201657 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-query-frontend-779849886d-hf7g4_2d2cf081-f13a-411b-bf24-5302a947dd9a/loki-query-frontend/0.log" Dec 01 20:51:54 crc kubenswrapper[4960]: I1201 20:51:54.592404 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vt7rt_b6b5eb9b-033d-4417-9289-953b2933520e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:54 crc kubenswrapper[4960]: I1201 20:51:54.732594 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-8bxh2_f45d70f4-68fe-491f-aa15-4e7d6a0a7812/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.028247 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-lokistack-querier-548665d79b-gd6lz_60517c3e-f430-453c-b6d4-3cdf124e100b/loki-querier/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.095626 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/init/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.394295 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/init/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.423014 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pbwwd_476f3fb9-9214-4448-9ff4-a5489883199b/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.478619 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-85f64749dc-jsbnv_fb9bb045-541f-477a-97fc-84e8bbd8f16b/dnsmasq-dns/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.635340 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a341d11-27da-4def-a478-c878a69e2e6b/glance-httpd/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.669927 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a341d11-27da-4def-a478-c878a69e2e6b/glance-log/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.815320 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1f3729c-7da0-4c16-b370-cacabf55f2b7/glance-httpd/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.827086 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a1f3729c-7da0-4c16-b370-cacabf55f2b7/glance-log/0.log" Dec 01 20:51:55 crc kubenswrapper[4960]: I1201 20:51:55.974811 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-4zdsr_cc7c39de-f6ea-4f00-ae86-a8f74ec35b71/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:56 crc kubenswrapper[4960]: I1201 20:51:56.146447 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-9bb2x_4aa6823d-0480-490a-977b-801110a10181/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:56 crc kubenswrapper[4960]: I1201 20:51:56.421034 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e118a262-1b53-4d2a-adce-3802b626cff1/kube-state-metrics/0.log" Dec 01 20:51:56 crc kubenswrapper[4960]: I1201 20:51:56.581705 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tvb6q_4c5eadde-8ce4-4009-9128-66d4d9657dfc/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:56 crc kubenswrapper[4960]: I1201 20:51:56.690970 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5f8c78b6fc-fdfln_987ef811-8e23-491e-98ff-64fcacab0f37/keystone-api/0.log" Dec 01 20:51:57 crc kubenswrapper[4960]: I1201 20:51:57.108372 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd769db7-hsw86_aee39569-b67b-419a-ad6a-461bbed0d585/neutron-httpd/0.log" Dec 01 20:51:57 crc kubenswrapper[4960]: I1201 20:51:57.156475 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd769db7-hsw86_aee39569-b67b-419a-ad6a-461bbed0d585/neutron-api/0.log" Dec 01 20:51:57 crc kubenswrapper[4960]: I1201 20:51:57.552771 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cloudkitty-proc-0_98bf64a0-1fed-4e0d-89ab-7c5b6e6fb98a/cloudkitty-proc/0.log" Dec 01 20:51:57 crc kubenswrapper[4960]: I1201 20:51:57.964782 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7qr58_6e32c90f-f0de-404d-be43-6953814aa12a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:58 crc kubenswrapper[4960]: I1201 20:51:58.468210 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3cb295cb-a487-406c-8f1d-517de7c1d245/nova-api-log/0.log" Dec 01 20:51:58 crc kubenswrapper[4960]: I1201 20:51:58.506061 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_878b6439-3b49-40d4-80e2-f01694590b67/nova-cell0-conductor-conductor/0.log" Dec 01 20:51:58 crc kubenswrapper[4960]: I1201 20:51:58.717441 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3cb295cb-a487-406c-8f1d-517de7c1d245/nova-api-api/0.log" Dec 01 20:51:59 crc kubenswrapper[4960]: I1201 20:51:59.510362 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_35daed02-01d5-4757-aa71-9c38e1397d19/nova-cell1-novncproxy-novncproxy/0.log" Dec 01 20:51:59 crc kubenswrapper[4960]: I1201 20:51:59.545135 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_38ad16c2-10a9-4a20-a070-e1ed80f62684/nova-cell1-conductor-conductor/0.log" Dec 01 20:51:59 crc kubenswrapper[4960]: I1201 20:51:59.561911 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-stff5_e37fd55a-a5e3-4844-93b3-a22664d77bcd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:51:59 crc kubenswrapper[4960]: I1201 20:51:59.782325 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_feac0db7-ea9a-4add-9c01-f942bc15b1bf/nova-metadata-log/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.058492 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_08cc519f-2c0d-444a-845a-9f403c9101be/nova-scheduler-scheduler/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.105527 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/mysql-bootstrap/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.327439 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/galera/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.390946 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2/mysql-bootstrap/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.588447 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/mysql-bootstrap/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.797776 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/galera/0.log" Dec 01 20:52:00 crc kubenswrapper[4960]: I1201 20:52:00.843500 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f9b50b6f-1af6-4283-8994-1cdc2b9e6c02/mysql-bootstrap/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.163859 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_5f778c19-cf9c-4c12-811d-9bc33aae5e90/openstackclient/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.243185 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_feac0db7-ea9a-4add-9c01-f942bc15b1bf/nova-metadata-metadata/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.244176 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-trsdq_5c1ff07c-096b-4952-a0cb-32fdcfaf154a/openstack-network-exporter/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.534540 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-nrf7z_87e16436-45f5-4392-b524-b86f1cc5c0b2/ovn-controller/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.580372 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server-init/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.738755 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server-init/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.884298 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovs-vswitchd/0.log" Dec 01 20:52:01 crc kubenswrapper[4960]: I1201 20:52:01.894258 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-9n6l4_9389e08d-0203-4f96-8278-fa1780a4201b/ovsdb-server/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.010932 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-bhl97_95e8c909-a7d7-4461-8fc7-c7644ad41000/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.132601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_07787692-57cc-473c-ab9b-276004fb9220/openstack-network-exporter/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.185189 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_07787692-57cc-473c-ab9b-276004fb9220/ovn-northd/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.356146 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a7eb168-4c2f-46c0-9292-de8e5c4383a5/ovsdbserver-nb/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.366625 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4a7eb168-4c2f-46c0-9292-de8e5c4383a5/openstack-network-exporter/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.589344 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1c6f94f2-a256-4e77-b6d1-39706b7c6ddc/openstack-network-exporter/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.613920 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1c6f94f2-a256-4e77-b6d1-39706b7c6ddc/ovsdbserver-sb/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.767371 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c7bf679b-vg6f9_f75be90e-8df9-4f2a-bbdb-6202c7b56cf9/placement-api/0.log" Dec 01 20:52:02 crc kubenswrapper[4960]: I1201 20:52:02.908860 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/init-config-reloader/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.050094 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-74c7bf679b-vg6f9_f75be90e-8df9-4f2a-bbdb-6202c7b56cf9/placement-log/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.240880 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/prometheus/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.246253 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/init-config-reloader/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.315596 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/config-reloader/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.344408 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_9a4f20da-f7ef-4710-9222-77baad783676/thanos-sidecar/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.527174 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/setup-container/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.822956 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/setup-container/0.log" Dec 01 20:52:03 crc kubenswrapper[4960]: I1201 20:52:03.997811 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/setup-container/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.030709 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_05a2ef84-5a97-4844-b037-2fcdf7fba8e4/rabbitmq/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.272040 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/setup-container/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.334001 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2f78cdc6-c6ed-4c26-9e66-f42bc510098d/rabbitmq/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.342751 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n57mc_1d22d084-b469-42e1-afdf-6c8ba0c4c59e/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.481980 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ndr5m_03306a2e-8c24-4479-bf7c-f6fbae3dd8ad/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.534704 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-r9h2n_bca07824-2e7a-4578-b833-02232c52ca0c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.797464 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-t8vj8_91e5e6e7-e948-43bb-9d94-c7cc7ed3ae34/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:04 crc kubenswrapper[4960]: I1201 20:52:04.871647 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mmbqh_9ab87b90-6085-4c16-a656-a6775aeb0b68/ssh-known-hosts-edpm-deployment/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.121239 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696549d5f9-6rmsc_c99a2254-ef5a-47fd-8ed6-bc98e9671077/proxy-server/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.163105 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-696549d5f9-6rmsc_c99a2254-ef5a-47fd-8ed6-bc98e9671077/proxy-httpd/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.361849 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gmqjb_aa3614b7-1a8b-4007-9e01-f6340ff9b0fe/swift-ring-rebalance/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.370192 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-auditor/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.437148 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-reaper/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.584024 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-server/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.598497 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/account-replicator/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.651158 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-auditor/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.744128 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-replicator/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.764500 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-server/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.800597 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/container-updater/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.941181 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-auditor/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.980883 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-expirer/0.log" Dec 01 20:52:05 crc kubenswrapper[4960]: I1201 20:52:05.984488 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-replicator/0.log" Dec 01 20:52:06 crc kubenswrapper[4960]: I1201 20:52:06.048685 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-server/0.log" Dec 01 20:52:06 crc kubenswrapper[4960]: I1201 20:52:06.869738 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/swift-recon-cron/0.log" Dec 01 20:52:06 crc kubenswrapper[4960]: I1201 20:52:06.965148 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/object-updater/0.log" Dec 01 20:52:07 crc kubenswrapper[4960]: I1201 20:52:07.049062 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_725698af-f23c-4262-96a4-2bfff12018fa/rsync/0.log" Dec 01 20:52:07 crc kubenswrapper[4960]: I1201 20:52:07.204154 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-9mvzc_00d8a61e-f3a9-4f45-9ff3-a196f814de12/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:07 crc kubenswrapper[4960]: I1201 20:52:07.269844 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d50fcead-df5c-40a0-b60f-4c195869572d/tempest-tests-tempest-tests-runner/0.log" Dec 01 20:52:07 crc kubenswrapper[4960]: I1201 20:52:07.435561 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_468c1f0b-62e1-492a-8496-288dc865ea98/test-operator-logs-container/0.log" Dec 01 20:52:07 crc kubenswrapper[4960]: I1201 20:52:07.563556 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-pwlj8_d8dcd369-d6cb-4472-8530-9c4d58ac2a0c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 01 20:52:12 crc kubenswrapper[4960]: I1201 20:52:12.144272 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6befffd7-b8c4-4c6c-900c-f6e1728f2a9e/memcached/0.log" Dec 01 20:52:13 crc kubenswrapper[4960]: I1201 20:52:13.870477 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:52:13 crc kubenswrapper[4960]: I1201 20:52:13.871664 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:52:13 crc kubenswrapper[4960]: I1201 20:52:13.871812 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:52:13 crc kubenswrapper[4960]: I1201 20:52:13.872544 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:52:13 crc kubenswrapper[4960]: I1201 20:52:13.873448 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786" gracePeriod=600 Dec 01 20:52:14 crc kubenswrapper[4960]: I1201 20:52:14.924964 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786" exitCode=0 Dec 01 20:52:14 crc kubenswrapper[4960]: I1201 20:52:14.925195 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786"} Dec 01 20:52:14 crc kubenswrapper[4960]: I1201 20:52:14.925563 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerStarted","Data":"7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57"} Dec 01 20:52:14 crc kubenswrapper[4960]: I1201 20:52:14.925585 4960 scope.go:117] "RemoveContainer" containerID="f32ff326c28e14fc50a7d87458e65bd879e0bc802a015f05ce88ec130f1609ed" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.110690 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.271649 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.283975 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.304151 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.500150 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/util/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.531853 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/pull/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.555150 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_2c09e5b1e9a9a4ca78abc65ab9e9961b7fb58875ca4cc4de57b3675e39sxsqz_87988f3c-1a59-4ebd-b80f-c590bd7e4e57/extract/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.687711 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8l7sf_238b3987-76a6-401d-a33c-5c2efd1bafcf/kube-rbac-proxy/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.804810 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-8l7sf_238b3987-76a6-401d-a33c-5c2efd1bafcf/manager/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.820150 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-p5qfh_cce3a29f-5414-4fee-a27c-9b6286930903/kube-rbac-proxy/0.log" Dec 01 20:52:38 crc kubenswrapper[4960]: I1201 20:52:38.956974 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-p5qfh_cce3a29f-5414-4fee-a27c-9b6286930903/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.002957 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6tbkt_a51010d5-7158-47ce-9428-25f5b2ca871f/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.042953 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-6tbkt_a51010d5-7158-47ce-9428-25f5b2ca871f/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.155395 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-pbxhm_018cf742-9a47-4355-879a-9e064f358748/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.278397 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-pbxhm_018cf742-9a47-4355-879a-9e064f358748/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.355131 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-6mqj7_0a07e8af-0ea2-4b3e-b9b7-b01f88025904/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.373179 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-6mqj7_0a07e8af-0ea2-4b3e-b9b7-b01f88025904/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.466418 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bsfbn_dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.556994 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-bsfbn_dfb2d8fc-0b91-4549-8cbf-04fbf79bf65d/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.620436 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pnrhx_b801b659-3926-4a9f-a36f-f752a6bd9563/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.825143 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zggh7_58815c45-1d27-4f86-9b6a-ffbb04c4e250/kube-rbac-proxy/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.872076 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-pnrhx_b801b659-3926-4a9f-a36f-f752a6bd9563/manager/0.log" Dec 01 20:52:39 crc kubenswrapper[4960]: I1201 20:52:39.910264 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-zggh7_58815c45-1d27-4f86-9b6a-ffbb04c4e250/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.011138 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-sfsvg_8b922299-0b5e-4595-bdab-1c759fd99465/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.098249 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-sfsvg_8b922299-0b5e-4595-bdab-1c759fd99465/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.154042 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-hcpbb_3b481993-f9fd-4d99-98e3-04b51c186e20/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.253355 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-hcpbb_3b481993-f9fd-4d99-98e3-04b51c186e20/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.316338 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-q8bv7_8c0036c6-9f32-4163-866d-c1f5786c5a2e/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.385345 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-q8bv7_8c0036c6-9f32-4163-866d-c1f5786c5a2e/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.494707 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-m24v2_de2501b9-bdd7-46c3-9128-20b2d9d5ab88/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.601590 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-m24v2_de2501b9-bdd7-46c3-9128-20b2d9d5ab88/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.701188 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-ld68p_0106714e-857b-4ade-9ee2-66fe8c12164c/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.774550 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-ld68p_0106714e-857b-4ade-9ee2-66fe8c12164c/manager/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.881965 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cjgcs_b869ba57-9812-4684-8cbb-48fc606c6aec/kube-rbac-proxy/0.log" Dec 01 20:52:40 crc kubenswrapper[4960]: I1201 20:52:40.936336 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-cjgcs_b869ba57-9812-4684-8cbb-48fc606c6aec/manager/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.056686 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6_6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c/kube-rbac-proxy/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.098123 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rrvv6_6994eaf1-2e28-45bb-8ae3-7c5bd2b51e9c/manager/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.363960 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2jktp_d683e2dd-4bba-4833-ae18-4636090ef1e5/registry-server/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.482493 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8455849bfc-vgkq8_9179ddbd-9397-405d-93d5-e32ce4963305/operator/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.610718 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-m9qpq_9b79fea5-fb83-4804-987d-b61e11999829/kube-rbac-proxy/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.694151 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vjb5b_b6018c51-4633-48dd-97ea-0d3d405f4e14/kube-rbac-proxy/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.743807 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-m9qpq_9b79fea5-fb83-4804-987d-b61e11999829/manager/0.log" Dec 01 20:52:41 crc kubenswrapper[4960]: I1201 20:52:41.876471 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-vjb5b_b6018c51-4633-48dd-97ea-0d3d405f4e14/manager/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.130338 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-5j8pz_123c59a4-e74e-4b51-9477-84d97ba60e36/operator/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.208307 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6zgq8_406103b8-0898-4d8f-badd-5f32648ad242/kube-rbac-proxy/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.291098 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6zgq8_406103b8-0898-4d8f-badd-5f32648ad242/manager/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.381896 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8466c66f58-t6nmf_d52ae874-2bed-4c1a-9c8c-a3e140971d4a/manager/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.410028 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-75c4b445dc-tfgqn_0b693fe0-7de6-4a0c-a931-21174d92c184/kube-rbac-proxy/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.585800 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rjb6d_2c77adc3-2d13-4c70-aefa-e4c238e61547/kube-rbac-proxy/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.640590 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rjb6d_2c77adc3-2d13-4c70-aefa-e4c238e61547/manager/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.728580 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-75c4b445dc-tfgqn_0b693fe0-7de6-4a0c-a931-21174d92c184/manager/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.766085 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k2567_82c72555-ba21-47f1-bfcd-ef954e788eb5/kube-rbac-proxy/0.log" Dec 01 20:52:42 crc kubenswrapper[4960]: I1201 20:52:42.779126 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-k2567_82c72555-ba21-47f1-bfcd-ef954e788eb5/manager/0.log" Dec 01 20:53:03 crc kubenswrapper[4960]: I1201 20:53:03.037883 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9qhvh_952f446b-a8d2-4887-bc67-cddfb9e20791/control-plane-machine-set-operator/0.log" Dec 01 20:53:03 crc kubenswrapper[4960]: I1201 20:53:03.207729 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mhbt7_b18c64d6-cd4c-4d07-9591-4c9868b33b2c/kube-rbac-proxy/0.log" Dec 01 20:53:03 crc kubenswrapper[4960]: I1201 20:53:03.219400 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mhbt7_b18c64d6-cd4c-4d07-9591-4c9868b33b2c/machine-api-operator/0.log" Dec 01 20:53:18 crc kubenswrapper[4960]: I1201 20:53:18.009525 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7dpzk_2e0d8acc-f304-45d4-b53c-dec0723c94e2/cert-manager-controller/0.log" Dec 01 20:53:18 crc kubenswrapper[4960]: I1201 20:53:18.116582 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-9lvxq_75151dc8-631d-42be-9b8c-e6b9a532c0da/cert-manager-cainjector/0.log" Dec 01 20:53:18 crc kubenswrapper[4960]: I1201 20:53:18.184620 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-fsw9v_12db57f7-a5f5-4a77-a412-d659e1490968/cert-manager-webhook/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.101748 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xpq6m_751bcd73-4c55-4fbb-a12d-0c9956a63e16/nmstate-console-plugin/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.300354 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-bwb2n_80d93577-3e89-47d7-b363-f22021509688/nmstate-handler/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.390452 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vh2zk_cba17f81-92f7-418b-a5c7-a184634c9acc/kube-rbac-proxy/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.400802 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-vh2zk_cba17f81-92f7-418b-a5c7-a184634c9acc/nmstate-metrics/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.533395 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-bvgrd_6377bee0-949f-4c1b-9758-c425ab4660bb/nmstate-operator/0.log" Dec 01 20:53:31 crc kubenswrapper[4960]: I1201 20:53:31.637384 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-5ndvb_138d040c-3866-4963-a2c8-26fb8a6c9811/nmstate-webhook/0.log" Dec 01 20:53:46 crc kubenswrapper[4960]: I1201 20:53:46.929804 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/kube-rbac-proxy/0.log" Dec 01 20:53:47 crc kubenswrapper[4960]: I1201 20:53:47.004671 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/manager/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.388300 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sm9mw_41410416-def6-44f6-9a25-14a6b9a50cec/kube-rbac-proxy/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.467859 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sm9mw_41410416-def6-44f6-9a25-14a6b9a50cec/controller/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.600400 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.807657 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.824255 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.824586 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:54:01 crc kubenswrapper[4960]: I1201 20:54:01.841986 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.047480 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.092462 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.104262 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.125541 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.278262 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-frr-files/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.281549 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-reloader/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.317488 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/cp-metrics/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.321525 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/controller/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.527999 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/kube-rbac-proxy/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.554601 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/frr-metrics/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.615573 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/kube-rbac-proxy-frr/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.803892 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/reloader/0.log" Dec 01 20:54:02 crc kubenswrapper[4960]: I1201 20:54:02.895863 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wjq2w_c7cef304-b739-45e6-b2c8-c729abfed45d/frr-k8s-webhook-server/0.log" Dec 01 20:54:03 crc kubenswrapper[4960]: I1201 20:54:03.079494 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-64f48564cc-pv444_8b72f468-b777-407c-8e36-30cc398644a6/manager/0.log" Dec 01 20:54:03 crc kubenswrapper[4960]: I1201 20:54:03.319659 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d7c4c664-7fttb_914a4b4c-d013-45ef-832d-35b600874887/webhook-server/0.log" Dec 01 20:54:03 crc kubenswrapper[4960]: I1201 20:54:03.394065 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qztxp_0c199599-aa8d-4e80-93f4-640a8ff141fa/kube-rbac-proxy/0.log" Dec 01 20:54:03 crc kubenswrapper[4960]: I1201 20:54:03.989169 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qztxp_0c199599-aa8d-4e80-93f4-640a8ff141fa/speaker/0.log" Dec 01 20:54:04 crc kubenswrapper[4960]: I1201 20:54:04.073729 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2dvm9_d98b4520-1781-455c-976c-259b9225fed7/frr/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.430723 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.590525 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.622136 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.663409 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.777532 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/util/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.810544 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/pull/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.813231 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frgt4n_2ff9d002-1e21-49cf-bced-1686ee42fe90/extract/0.log" Dec 01 20:54:18 crc kubenswrapper[4960]: I1201 20:54:18.967337 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.133094 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.150968 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.165497 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.356530 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/pull/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.362011 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/util/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.402527 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f835gkln_96e5ebac-e485-4307-a5ff-73e77e788e02/extract/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.529963 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.704684 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.711335 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.735785 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.878329 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-content/0.log" Dec 01 20:54:19 crc kubenswrapper[4960]: I1201 20:54:19.911281 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/extract-utilities/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.054675 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.361255 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.374403 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.391222 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcfpr_eb389473-d315-4d60-830b-a0d513d4e07b/registry-server/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.405484 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.504027 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-utilities/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.563898 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/extract-content/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.716587 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qsjbd_a968fcf3-9935-4cfe-abdc-82c62d9b3b12/marketplace-operator/0.log" Dec 01 20:54:20 crc kubenswrapper[4960]: I1201 20:54:20.897460 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.059038 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.125899 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.145283 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8v7bs_bd9c07a1-998c-479c-8440-2d7a4b27bb2a/registry-server/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.189673 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.353910 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.360706 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/extract-utilities/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.479214 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-k4nsz_ccd8de2a-98a4-4403-9f25-3bb5b7e8d765/registry-server/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.572593 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.705078 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.736769 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.749536 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.923704 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-content/0.log" Dec 01 20:54:21 crc kubenswrapper[4960]: I1201 20:54:21.924255 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/extract-utilities/0.log" Dec 01 20:54:22 crc kubenswrapper[4960]: I1201 20:54:22.413036 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c4n7v_81be1171-9e3d-49fd-924e-044b7d7c5023/registry-server/0.log" Dec 01 20:54:35 crc kubenswrapper[4960]: I1201 20:54:35.720822 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-dhxwf_ba9ca50d-24cd-4d42-b130-25c0e744bcb4/prometheus-operator/0.log" Dec 01 20:54:35 crc kubenswrapper[4960]: I1201 20:54:35.877935 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-585b7cbc88-lspmg_d6e2b5c4-73c4-45a8-8d81-08adbe28e39f/prometheus-operator-admission-webhook/0.log" Dec 01 20:54:35 crc kubenswrapper[4960]: I1201 20:54:35.902135 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-585b7cbc88-smxth_069dc362-519d-4349-b53e-288ee73b9916/prometheus-operator-admission-webhook/0.log" Dec 01 20:54:36 crc kubenswrapper[4960]: I1201 20:54:36.109949 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-gfmgc_953224f9-3cdd-4e9c-98ac-aec3fc6ccc23/operator/0.log" Dec 01 20:54:36 crc kubenswrapper[4960]: I1201 20:54:36.143144 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-tktnr_4f86337f-46f5-46aa-aaba-41179da85cab/perses-operator/0.log" Dec 01 20:54:43 crc kubenswrapper[4960]: I1201 20:54:43.870618 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:54:43 crc kubenswrapper[4960]: I1201 20:54:43.871189 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:54:50 crc kubenswrapper[4960]: I1201 20:54:50.394633 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/manager/0.log" Dec 01 20:54:50 crc kubenswrapper[4960]: I1201 20:54:50.415399 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-6dfbdd7cb8-tcjcb_40cc1143-fbb0-4ef7-bb03-a9ef3b1c986b/kube-rbac-proxy/0.log" Dec 01 20:55:13 crc kubenswrapper[4960]: I1201 20:55:13.870767 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:55:13 crc kubenswrapper[4960]: I1201 20:55:13.871464 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:55:14 crc kubenswrapper[4960]: E1201 20:55:14.735725 4960 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.177:59114->38.102.83.177:40975: write tcp 38.102.83.177:59114->38.102.83.177:40975: write: broken pipe Dec 01 20:55:43 crc kubenswrapper[4960]: I1201 20:55:43.870674 4960 patch_prober.go:28] interesting pod/machine-config-daemon-ct7db container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 01 20:55:43 crc kubenswrapper[4960]: I1201 20:55:43.871302 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 01 20:55:43 crc kubenswrapper[4960]: I1201 20:55:43.871348 4960 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" Dec 01 20:55:43 crc kubenswrapper[4960]: I1201 20:55:43.872324 4960 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57"} pod="openshift-machine-config-operator/machine-config-daemon-ct7db" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 01 20:55:43 crc kubenswrapper[4960]: I1201 20:55:43.872379 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerName="machine-config-daemon" containerID="cri-o://7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" gracePeriod=600 Dec 01 20:55:44 crc kubenswrapper[4960]: I1201 20:55:44.100688 4960 generic.go:334] "Generic (PLEG): container finished" podID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" exitCode=0 Dec 01 20:55:44 crc kubenswrapper[4960]: I1201 20:55:44.100739 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" event={"ID":"b6dbabf7-fd52-4f8d-9bca-093018d1c0b9","Type":"ContainerDied","Data":"7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57"} Dec 01 20:55:44 crc kubenswrapper[4960]: I1201 20:55:44.100782 4960 scope.go:117] "RemoveContainer" containerID="03c76c359f07aa8153a64cb0785cd47f6681d0a7dbaf7ed02e152a451332c786" Dec 01 20:55:44 crc kubenswrapper[4960]: E1201 20:55:44.500321 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:55:45 crc kubenswrapper[4960]: I1201 20:55:45.113194 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:55:45 crc kubenswrapper[4960]: E1201 20:55:45.113528 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:55:57 crc kubenswrapper[4960]: I1201 20:55:57.324444 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:55:57 crc kubenswrapper[4960]: E1201 20:55:57.325342 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:56:11 crc kubenswrapper[4960]: I1201 20:56:11.329172 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:56:11 crc kubenswrapper[4960]: E1201 20:56:11.329878 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:56:25 crc kubenswrapper[4960]: I1201 20:56:25.330703 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:56:25 crc kubenswrapper[4960]: E1201 20:56:25.332775 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:56:34 crc kubenswrapper[4960]: I1201 20:56:34.699465 4960 generic.go:334] "Generic (PLEG): container finished" podID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerID="a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c" exitCode=0 Dec 01 20:56:34 crc kubenswrapper[4960]: I1201 20:56:34.699586 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" event={"ID":"31fa442c-a776-4f31-9a4b-44ac89b0ff31","Type":"ContainerDied","Data":"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c"} Dec 01 20:56:34 crc kubenswrapper[4960]: I1201 20:56:34.700844 4960 scope.go:117] "RemoveContainer" containerID="a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c" Dec 01 20:56:34 crc kubenswrapper[4960]: I1201 20:56:34.989582 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ztv6_must-gather-pjwp5_31fa442c-a776-4f31-9a4b-44ac89b0ff31/gather/0.log" Dec 01 20:56:40 crc kubenswrapper[4960]: I1201 20:56:40.325797 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:56:40 crc kubenswrapper[4960]: E1201 20:56:40.326950 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:56:46 crc kubenswrapper[4960]: I1201 20:56:46.989345 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ztv6/must-gather-pjwp5"] Dec 01 20:56:46 crc kubenswrapper[4960]: I1201 20:56:46.990220 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="copy" containerID="cri-o://2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6" gracePeriod=2 Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.000798 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ztv6/must-gather-pjwp5"] Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.541841 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ztv6_must-gather-pjwp5_31fa442c-a776-4f31-9a4b-44ac89b0ff31/copy/0.log" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.542552 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.638625 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvj27\" (UniqueName: \"kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27\") pod \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.638724 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output\") pod \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\" (UID: \"31fa442c-a776-4f31-9a4b-44ac89b0ff31\") " Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.655917 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27" (OuterVolumeSpecName: "kube-api-access-rvj27") pod "31fa442c-a776-4f31-9a4b-44ac89b0ff31" (UID: "31fa442c-a776-4f31-9a4b-44ac89b0ff31"). InnerVolumeSpecName "kube-api-access-rvj27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.743785 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvj27\" (UniqueName: \"kubernetes.io/projected/31fa442c-a776-4f31-9a4b-44ac89b0ff31-kube-api-access-rvj27\") on node \"crc\" DevicePath \"\"" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.809458 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "31fa442c-a776-4f31-9a4b-44ac89b0ff31" (UID: "31fa442c-a776-4f31-9a4b-44ac89b0ff31"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.845868 4960 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31fa442c-a776-4f31-9a4b-44ac89b0ff31-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.877992 4960 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ztv6_must-gather-pjwp5_31fa442c-a776-4f31-9a4b-44ac89b0ff31/copy/0.log" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.878379 4960 generic.go:334] "Generic (PLEG): container finished" podID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerID="2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6" exitCode=143 Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.878463 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ztv6/must-gather-pjwp5" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.878426 4960 scope.go:117] "RemoveContainer" containerID="2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.902331 4960 scope.go:117] "RemoveContainer" containerID="a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.951454 4960 scope.go:117] "RemoveContainer" containerID="2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6" Dec 01 20:56:47 crc kubenswrapper[4960]: E1201 20:56:47.951912 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6\": container with ID starting with 2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6 not found: ID does not exist" containerID="2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.951948 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6"} err="failed to get container status \"2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6\": rpc error: code = NotFound desc = could not find container \"2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6\": container with ID starting with 2d7aa73a4bb24813411d4fb6c0ff380eb8999882946b6d650d27164009e3fbb6 not found: ID does not exist" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.951973 4960 scope.go:117] "RemoveContainer" containerID="a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c" Dec 01 20:56:47 crc kubenswrapper[4960]: E1201 20:56:47.952428 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c\": container with ID starting with a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c not found: ID does not exist" containerID="a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c" Dec 01 20:56:47 crc kubenswrapper[4960]: I1201 20:56:47.952471 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c"} err="failed to get container status \"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c\": rpc error: code = NotFound desc = could not find container \"a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c\": container with ID starting with a577588c8e2f92b962f51323efae02e94c0f77ba28850b18b309d5b730b8301c not found: ID does not exist" Dec 01 20:56:49 crc kubenswrapper[4960]: I1201 20:56:49.335820 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" path="/var/lib/kubelet/pods/31fa442c-a776-4f31-9a4b-44ac89b0ff31/volumes" Dec 01 20:56:52 crc kubenswrapper[4960]: I1201 20:56:52.324495 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:56:52 crc kubenswrapper[4960]: E1201 20:56:52.325103 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.277091 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:56:53 crc kubenswrapper[4960]: E1201 20:56:53.277749 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3cb255-1f7e-4256-bbf3-80b208a265db" containerName="container-00" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.277771 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3cb255-1f7e-4256-bbf3-80b208a265db" containerName="container-00" Dec 01 20:56:53 crc kubenswrapper[4960]: E1201 20:56:53.277807 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="gather" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.277816 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="gather" Dec 01 20:56:53 crc kubenswrapper[4960]: E1201 20:56:53.277852 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="copy" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.277862 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="copy" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.278161 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="copy" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.278186 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3cb255-1f7e-4256-bbf3-80b208a265db" containerName="container-00" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.278231 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="31fa442c-a776-4f31-9a4b-44ac89b0ff31" containerName="gather" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.280452 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.294396 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.352502 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.352632 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7zj2\" (UniqueName: \"kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.352667 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.455274 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.455688 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7zj2\" (UniqueName: \"kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.455712 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.455761 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.456590 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.477919 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7zj2\" (UniqueName: \"kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2\") pod \"redhat-marketplace-7lh7b\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:53 crc kubenswrapper[4960]: I1201 20:56:53.606153 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:56:54 crc kubenswrapper[4960]: I1201 20:56:54.092316 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:56:54 crc kubenswrapper[4960]: I1201 20:56:54.978170 4960 generic.go:334] "Generic (PLEG): container finished" podID="2106a183-8e68-42ea-87d4-1465649b7385" containerID="90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f" exitCode=0 Dec 01 20:56:54 crc kubenswrapper[4960]: I1201 20:56:54.978558 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerDied","Data":"90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f"} Dec 01 20:56:54 crc kubenswrapper[4960]: I1201 20:56:54.979277 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerStarted","Data":"a5b9d66cfb04d4d756ecf13a966582f02c3bcef6806cece18c195a24832f2b10"} Dec 01 20:56:54 crc kubenswrapper[4960]: I1201 20:56:54.982199 4960 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 01 20:56:55 crc kubenswrapper[4960]: I1201 20:56:55.989736 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerStarted","Data":"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386"} Dec 01 20:56:57 crc kubenswrapper[4960]: I1201 20:56:57.001062 4960 generic.go:334] "Generic (PLEG): container finished" podID="2106a183-8e68-42ea-87d4-1465649b7385" containerID="ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386" exitCode=0 Dec 01 20:56:57 crc kubenswrapper[4960]: I1201 20:56:57.001460 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerDied","Data":"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386"} Dec 01 20:56:59 crc kubenswrapper[4960]: I1201 20:56:59.045068 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerStarted","Data":"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438"} Dec 01 20:56:59 crc kubenswrapper[4960]: I1201 20:56:59.075131 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7lh7b" podStartSLOduration=3.380571207 podStartE2EDuration="6.075091994s" podCreationTimestamp="2025-12-01 20:56:53 +0000 UTC" firstStartedPulling="2025-12-01 20:56:54.98197593 +0000 UTC m=+4650.269467599" lastFinishedPulling="2025-12-01 20:56:57.676496717 +0000 UTC m=+4652.963988386" observedRunningTime="2025-12-01 20:56:59.070002375 +0000 UTC m=+4654.357494054" watchObservedRunningTime="2025-12-01 20:56:59.075091994 +0000 UTC m=+4654.362583663" Dec 01 20:57:03 crc kubenswrapper[4960]: I1201 20:57:03.607011 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:03 crc kubenswrapper[4960]: I1201 20:57:03.607474 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:03 crc kubenswrapper[4960]: I1201 20:57:03.654048 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:04 crc kubenswrapper[4960]: I1201 20:57:04.171281 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:04 crc kubenswrapper[4960]: I1201 20:57:04.253785 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:57:06 crc kubenswrapper[4960]: I1201 20:57:06.133876 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7lh7b" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="registry-server" containerID="cri-o://30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438" gracePeriod=2 Dec 01 20:57:06 crc kubenswrapper[4960]: I1201 20:57:06.933520 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-696549d5f9-6rmsc" podUID="c99a2254-ef5a-47fd-8ed6-bc98e9671077" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.001641 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.143989 4960 generic.go:334] "Generic (PLEG): container finished" podID="2106a183-8e68-42ea-87d4-1465649b7385" containerID="30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438" exitCode=0 Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.144220 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerDied","Data":"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438"} Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.145137 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7lh7b" event={"ID":"2106a183-8e68-42ea-87d4-1465649b7385","Type":"ContainerDied","Data":"a5b9d66cfb04d4d756ecf13a966582f02c3bcef6806cece18c195a24832f2b10"} Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.144470 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7zj2\" (UniqueName: \"kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2\") pod \"2106a183-8e68-42ea-87d4-1465649b7385\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.144293 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7lh7b" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.145177 4960 scope.go:117] "RemoveContainer" containerID="30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.145699 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content\") pod \"2106a183-8e68-42ea-87d4-1465649b7385\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.145734 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities\") pod \"2106a183-8e68-42ea-87d4-1465649b7385\" (UID: \"2106a183-8e68-42ea-87d4-1465649b7385\") " Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.146794 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities" (OuterVolumeSpecName: "utilities") pod "2106a183-8e68-42ea-87d4-1465649b7385" (UID: "2106a183-8e68-42ea-87d4-1465649b7385"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.150171 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2" (OuterVolumeSpecName: "kube-api-access-b7zj2") pod "2106a183-8e68-42ea-87d4-1465649b7385" (UID: "2106a183-8e68-42ea-87d4-1465649b7385"). InnerVolumeSpecName "kube-api-access-b7zj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.166218 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2106a183-8e68-42ea-87d4-1465649b7385" (UID: "2106a183-8e68-42ea-87d4-1465649b7385"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.245352 4960 scope.go:117] "RemoveContainer" containerID="ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.247925 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7zj2\" (UniqueName: \"kubernetes.io/projected/2106a183-8e68-42ea-87d4-1465649b7385-kube-api-access-b7zj2\") on node \"crc\" DevicePath \"\"" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.247957 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.247969 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2106a183-8e68-42ea-87d4-1465649b7385-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.265516 4960 scope.go:117] "RemoveContainer" containerID="90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.324723 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:57:07 crc kubenswrapper[4960]: E1201 20:57:07.325242 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.469945 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.479067 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7lh7b"] Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.504222 4960 scope.go:117] "RemoveContainer" containerID="30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438" Dec 01 20:57:07 crc kubenswrapper[4960]: E1201 20:57:07.504738 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438\": container with ID starting with 30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438 not found: ID does not exist" containerID="30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.504865 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438"} err="failed to get container status \"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438\": rpc error: code = NotFound desc = could not find container \"30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438\": container with ID starting with 30a9b12e61b1b97b651728321772c737c8d23f19a64b2f13a190a9e572583438 not found: ID does not exist" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.504950 4960 scope.go:117] "RemoveContainer" containerID="ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386" Dec 01 20:57:07 crc kubenswrapper[4960]: E1201 20:57:07.505433 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386\": container with ID starting with ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386 not found: ID does not exist" containerID="ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.505473 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386"} err="failed to get container status \"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386\": rpc error: code = NotFound desc = could not find container \"ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386\": container with ID starting with ec385aa978aa1f3a6452cd8d4d70e14b5739f3be15e894af3cb820cd5a20f386 not found: ID does not exist" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.505499 4960 scope.go:117] "RemoveContainer" containerID="90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f" Dec 01 20:57:07 crc kubenswrapper[4960]: E1201 20:57:07.505785 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f\": container with ID starting with 90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f not found: ID does not exist" containerID="90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f" Dec 01 20:57:07 crc kubenswrapper[4960]: I1201 20:57:07.505815 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f"} err="failed to get container status \"90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f\": rpc error: code = NotFound desc = could not find container \"90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f\": container with ID starting with 90c9fcc67b5e58b646d7eacf21080ad03e96c149dcea9de619720a7fab46eb7f not found: ID does not exist" Dec 01 20:57:09 crc kubenswrapper[4960]: I1201 20:57:09.362408 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2106a183-8e68-42ea-87d4-1465649b7385" path="/var/lib/kubelet/pods/2106a183-8e68-42ea-87d4-1465649b7385/volumes" Dec 01 20:57:16 crc kubenswrapper[4960]: I1201 20:57:16.944361 4960 scope.go:117] "RemoveContainer" containerID="f15ea7ebc9caa9551247852151f9d28cf3dce6ad5a15f89da6fc34c4b19992ff" Dec 01 20:57:17 crc kubenswrapper[4960]: I1201 20:57:17.006765 4960 scope.go:117] "RemoveContainer" containerID="8b3ad2b57d97604a86c0af119eaabba75e7a50f3b9d24584baaba18c0c3c7c41" Dec 01 20:57:21 crc kubenswrapper[4960]: I1201 20:57:21.324246 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:57:21 crc kubenswrapper[4960]: E1201 20:57:21.324961 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:57:35 crc kubenswrapper[4960]: I1201 20:57:35.336696 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:57:35 crc kubenswrapper[4960]: E1201 20:57:35.337632 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:57:47 crc kubenswrapper[4960]: I1201 20:57:47.324632 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:57:47 crc kubenswrapper[4960]: E1201 20:57:47.325479 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.562332 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:57:54 crc kubenswrapper[4960]: E1201 20:57:54.563393 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="registry-server" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.563411 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="registry-server" Dec 01 20:57:54 crc kubenswrapper[4960]: E1201 20:57:54.563454 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="extract-utilities" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.563462 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="extract-utilities" Dec 01 20:57:54 crc kubenswrapper[4960]: E1201 20:57:54.563475 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="extract-content" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.563482 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="extract-content" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.563744 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="2106a183-8e68-42ea-87d4-1465649b7385" containerName="registry-server" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.565647 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.581585 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.698428 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25cpm\" (UniqueName: \"kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.698517 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.698683 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.800695 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25cpm\" (UniqueName: \"kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.800783 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.800872 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.801200 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.801734 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.819979 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25cpm\" (UniqueName: \"kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm\") pod \"redhat-operators-j6dv8\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:54 crc kubenswrapper[4960]: I1201 20:57:54.893380 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:57:55 crc kubenswrapper[4960]: I1201 20:57:55.390184 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:57:55 crc kubenswrapper[4960]: I1201 20:57:55.657081 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerStarted","Data":"34d7e26b550cb3ee2fddebef95ee9739d1420158364058929ee2dd7937a66b36"} Dec 01 20:57:56 crc kubenswrapper[4960]: I1201 20:57:56.669406 4960 generic.go:334] "Generic (PLEG): container finished" podID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerID="eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19" exitCode=0 Dec 01 20:57:56 crc kubenswrapper[4960]: I1201 20:57:56.669518 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerDied","Data":"eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19"} Dec 01 20:57:58 crc kubenswrapper[4960]: I1201 20:57:58.951358 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:57:58 crc kubenswrapper[4960]: I1201 20:57:58.955620 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:58 crc kubenswrapper[4960]: I1201 20:57:58.965033 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.095640 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.095884 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttbp4\" (UniqueName: \"kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.096196 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.198633 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.198771 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.198889 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttbp4\" (UniqueName: \"kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.199268 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.199286 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.217948 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttbp4\" (UniqueName: \"kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4\") pod \"community-operators-cn9mp\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.280590 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.698462 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerStarted","Data":"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b"} Dec 01 20:57:59 crc kubenswrapper[4960]: W1201 20:57:59.891857 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee493fec_2662_4f83_969e_fdaa2e4f2bac.slice/crio-04bd965c415c7f7b3609911830dc410ce9f5cf50ee7b43357b01f4f83f4f8fd6 WatchSource:0}: Error finding container 04bd965c415c7f7b3609911830dc410ce9f5cf50ee7b43357b01f4f83f4f8fd6: Status 404 returned error can't find the container with id 04bd965c415c7f7b3609911830dc410ce9f5cf50ee7b43357b01f4f83f4f8fd6 Dec 01 20:57:59 crc kubenswrapper[4960]: I1201 20:57:59.893134 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:58:00 crc kubenswrapper[4960]: I1201 20:58:00.710603 4960 generic.go:334] "Generic (PLEG): container finished" podID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerID="f1c4d7584c13c1cf7882bf26b6f78b62997e1098b877e0f33719463789a70bd6" exitCode=0 Dec 01 20:58:00 crc kubenswrapper[4960]: I1201 20:58:00.710678 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerDied","Data":"f1c4d7584c13c1cf7882bf26b6f78b62997e1098b877e0f33719463789a70bd6"} Dec 01 20:58:00 crc kubenswrapper[4960]: I1201 20:58:00.710984 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerStarted","Data":"04bd965c415c7f7b3609911830dc410ce9f5cf50ee7b43357b01f4f83f4f8fd6"} Dec 01 20:58:01 crc kubenswrapper[4960]: I1201 20:58:01.324676 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:58:01 crc kubenswrapper[4960]: E1201 20:58:01.325188 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:58:01 crc kubenswrapper[4960]: I1201 20:58:01.725147 4960 generic.go:334] "Generic (PLEG): container finished" podID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerID="34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b" exitCode=0 Dec 01 20:58:01 crc kubenswrapper[4960]: I1201 20:58:01.725240 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerDied","Data":"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b"} Dec 01 20:58:01 crc kubenswrapper[4960]: I1201 20:58:01.732373 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerStarted","Data":"a97dfc643999295423b8bf0c85261d6854f94c4b7e0a993b3b05ff2c85a38ecb"} Dec 01 20:58:02 crc kubenswrapper[4960]: I1201 20:58:02.742461 4960 generic.go:334] "Generic (PLEG): container finished" podID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerID="a97dfc643999295423b8bf0c85261d6854f94c4b7e0a993b3b05ff2c85a38ecb" exitCode=0 Dec 01 20:58:02 crc kubenswrapper[4960]: I1201 20:58:02.742506 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerDied","Data":"a97dfc643999295423b8bf0c85261d6854f94c4b7e0a993b3b05ff2c85a38ecb"} Dec 01 20:58:02 crc kubenswrapper[4960]: I1201 20:58:02.746499 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerStarted","Data":"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf"} Dec 01 20:58:02 crc kubenswrapper[4960]: I1201 20:58:02.791805 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j6dv8" podStartSLOduration=3.19710196 podStartE2EDuration="8.791783952s" podCreationTimestamp="2025-12-01 20:57:54 +0000 UTC" firstStartedPulling="2025-12-01 20:57:56.671464671 +0000 UTC m=+4711.958956340" lastFinishedPulling="2025-12-01 20:58:02.266146643 +0000 UTC m=+4717.553638332" observedRunningTime="2025-12-01 20:58:02.783705919 +0000 UTC m=+4718.071197628" watchObservedRunningTime="2025-12-01 20:58:02.791783952 +0000 UTC m=+4718.079275641" Dec 01 20:58:03 crc kubenswrapper[4960]: I1201 20:58:03.759273 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerStarted","Data":"cd4e15107ada08f8a2134f03d77cc3ac8eef5f21383a18091648a3817fe63709"} Dec 01 20:58:03 crc kubenswrapper[4960]: I1201 20:58:03.777151 4960 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cn9mp" podStartSLOduration=3.240437765 podStartE2EDuration="5.777132774s" podCreationTimestamp="2025-12-01 20:57:58 +0000 UTC" firstStartedPulling="2025-12-01 20:58:00.712158563 +0000 UTC m=+4715.999650232" lastFinishedPulling="2025-12-01 20:58:03.248853572 +0000 UTC m=+4718.536345241" observedRunningTime="2025-12-01 20:58:03.774305045 +0000 UTC m=+4719.061796714" watchObservedRunningTime="2025-12-01 20:58:03.777132774 +0000 UTC m=+4719.064624443" Dec 01 20:58:04 crc kubenswrapper[4960]: I1201 20:58:04.893658 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:04 crc kubenswrapper[4960]: I1201 20:58:04.894099 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:05 crc kubenswrapper[4960]: I1201 20:58:05.963524 4960 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j6dv8" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="registry-server" probeResult="failure" output=< Dec 01 20:58:05 crc kubenswrapper[4960]: timeout: failed to connect service ":50051" within 1s Dec 01 20:58:05 crc kubenswrapper[4960]: > Dec 01 20:58:09 crc kubenswrapper[4960]: I1201 20:58:09.281008 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:09 crc kubenswrapper[4960]: I1201 20:58:09.281332 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:09 crc kubenswrapper[4960]: I1201 20:58:09.342402 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:09 crc kubenswrapper[4960]: I1201 20:58:09.873659 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:09 crc kubenswrapper[4960]: I1201 20:58:09.933021 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:58:11 crc kubenswrapper[4960]: I1201 20:58:11.839921 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cn9mp" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="registry-server" containerID="cri-o://cd4e15107ada08f8a2134f03d77cc3ac8eef5f21383a18091648a3817fe63709" gracePeriod=2 Dec 01 20:58:12 crc kubenswrapper[4960]: I1201 20:58:12.851671 4960 generic.go:334] "Generic (PLEG): container finished" podID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerID="cd4e15107ada08f8a2134f03d77cc3ac8eef5f21383a18091648a3817fe63709" exitCode=0 Dec 01 20:58:12 crc kubenswrapper[4960]: I1201 20:58:12.851758 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerDied","Data":"cd4e15107ada08f8a2134f03d77cc3ac8eef5f21383a18091648a3817fe63709"} Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.722694 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.845937 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities\") pod \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.846172 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content\") pod \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.846331 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttbp4\" (UniqueName: \"kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4\") pod \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\" (UID: \"ee493fec-2662-4f83-969e-fdaa2e4f2bac\") " Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.847187 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities" (OuterVolumeSpecName: "utilities") pod "ee493fec-2662-4f83-969e-fdaa2e4f2bac" (UID: "ee493fec-2662-4f83-969e-fdaa2e4f2bac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.858688 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4" (OuterVolumeSpecName: "kube-api-access-ttbp4") pod "ee493fec-2662-4f83-969e-fdaa2e4f2bac" (UID: "ee493fec-2662-4f83-969e-fdaa2e4f2bac"). InnerVolumeSpecName "kube-api-access-ttbp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.880691 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cn9mp" event={"ID":"ee493fec-2662-4f83-969e-fdaa2e4f2bac","Type":"ContainerDied","Data":"04bd965c415c7f7b3609911830dc410ce9f5cf50ee7b43357b01f4f83f4f8fd6"} Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.880911 4960 scope.go:117] "RemoveContainer" containerID="cd4e15107ada08f8a2134f03d77cc3ac8eef5f21383a18091648a3817fe63709" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.881122 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cn9mp" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.911200 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee493fec-2662-4f83-969e-fdaa2e4f2bac" (UID: "ee493fec-2662-4f83-969e-fdaa2e4f2bac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.928755 4960 scope.go:117] "RemoveContainer" containerID="a97dfc643999295423b8bf0c85261d6854f94c4b7e0a993b3b05ff2c85a38ecb" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.948304 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.948349 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttbp4\" (UniqueName: \"kubernetes.io/projected/ee493fec-2662-4f83-969e-fdaa2e4f2bac-kube-api-access-ttbp4\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.948362 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee493fec-2662-4f83-969e-fdaa2e4f2bac-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:13 crc kubenswrapper[4960]: I1201 20:58:13.958848 4960 scope.go:117] "RemoveContainer" containerID="f1c4d7584c13c1cf7882bf26b6f78b62997e1098b877e0f33719463789a70bd6" Dec 01 20:58:14 crc kubenswrapper[4960]: I1201 20:58:14.224322 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:58:14 crc kubenswrapper[4960]: I1201 20:58:14.234350 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cn9mp"] Dec 01 20:58:14 crc kubenswrapper[4960]: I1201 20:58:14.951928 4960 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:15 crc kubenswrapper[4960]: I1201 20:58:15.130770 4960 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:15 crc kubenswrapper[4960]: I1201 20:58:15.341229 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" path="/var/lib/kubelet/pods/ee493fec-2662-4f83-969e-fdaa2e4f2bac/volumes" Dec 01 20:58:15 crc kubenswrapper[4960]: I1201 20:58:15.963551 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:58:16 crc kubenswrapper[4960]: I1201 20:58:16.324429 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:58:16 crc kubenswrapper[4960]: E1201 20:58:16.324690 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:58:16 crc kubenswrapper[4960]: I1201 20:58:16.919928 4960 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j6dv8" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="registry-server" containerID="cri-o://a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf" gracePeriod=2 Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.504374 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.641359 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content\") pod \"702ffbff-4550-40d7-86fa-3fb78b676fce\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.641429 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25cpm\" (UniqueName: \"kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm\") pod \"702ffbff-4550-40d7-86fa-3fb78b676fce\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.641544 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities\") pod \"702ffbff-4550-40d7-86fa-3fb78b676fce\" (UID: \"702ffbff-4550-40d7-86fa-3fb78b676fce\") " Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.642035 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities" (OuterVolumeSpecName: "utilities") pod "702ffbff-4550-40d7-86fa-3fb78b676fce" (UID: "702ffbff-4550-40d7-86fa-3fb78b676fce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.642178 4960 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-utilities\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.649515 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm" (OuterVolumeSpecName: "kube-api-access-25cpm") pod "702ffbff-4550-40d7-86fa-3fb78b676fce" (UID: "702ffbff-4550-40d7-86fa-3fb78b676fce"). InnerVolumeSpecName "kube-api-access-25cpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.744952 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25cpm\" (UniqueName: \"kubernetes.io/projected/702ffbff-4550-40d7-86fa-3fb78b676fce-kube-api-access-25cpm\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.746226 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "702ffbff-4550-40d7-86fa-3fb78b676fce" (UID: "702ffbff-4550-40d7-86fa-3fb78b676fce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.846236 4960 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702ffbff-4550-40d7-86fa-3fb78b676fce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.934175 4960 generic.go:334] "Generic (PLEG): container finished" podID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerID="a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf" exitCode=0 Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.934222 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6dv8" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.934246 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerDied","Data":"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf"} Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.934302 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6dv8" event={"ID":"702ffbff-4550-40d7-86fa-3fb78b676fce","Type":"ContainerDied","Data":"34d7e26b550cb3ee2fddebef95ee9739d1420158364058929ee2dd7937a66b36"} Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.934337 4960 scope.go:117] "RemoveContainer" containerID="a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.958422 4960 scope.go:117] "RemoveContainer" containerID="34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b" Dec 01 20:58:17 crc kubenswrapper[4960]: I1201 20:58:17.986649 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.001398 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j6dv8"] Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.010108 4960 scope.go:117] "RemoveContainer" containerID="eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.076338 4960 scope.go:117] "RemoveContainer" containerID="a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf" Dec 01 20:58:18 crc kubenswrapper[4960]: E1201 20:58:18.076984 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf\": container with ID starting with a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf not found: ID does not exist" containerID="a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.077036 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf"} err="failed to get container status \"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf\": rpc error: code = NotFound desc = could not find container \"a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf\": container with ID starting with a625af192f2e19793763b612933467035f9f415490e2116857700df86c241eaf not found: ID does not exist" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.077067 4960 scope.go:117] "RemoveContainer" containerID="34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b" Dec 01 20:58:18 crc kubenswrapper[4960]: E1201 20:58:18.077570 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b\": container with ID starting with 34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b not found: ID does not exist" containerID="34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.077611 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b"} err="failed to get container status \"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b\": rpc error: code = NotFound desc = could not find container \"34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b\": container with ID starting with 34c904c6c9737bc50086d8f78fe741a06e2e5ea07582cbd394947e8bd919c98b not found: ID does not exist" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.077636 4960 scope.go:117] "RemoveContainer" containerID="eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19" Dec 01 20:58:18 crc kubenswrapper[4960]: E1201 20:58:18.077993 4960 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19\": container with ID starting with eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19 not found: ID does not exist" containerID="eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19" Dec 01 20:58:18 crc kubenswrapper[4960]: I1201 20:58:18.078028 4960 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19"} err="failed to get container status \"eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19\": rpc error: code = NotFound desc = could not find container \"eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19\": container with ID starting with eaa2db9f9f0751b5c66e0fd949bd55a7fb7c96ed34d690954f617ec48cc91c19 not found: ID does not exist" Dec 01 20:58:19 crc kubenswrapper[4960]: I1201 20:58:19.341385 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" path="/var/lib/kubelet/pods/702ffbff-4550-40d7-86fa-3fb78b676fce/volumes" Dec 01 20:58:29 crc kubenswrapper[4960]: I1201 20:58:29.325835 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:58:29 crc kubenswrapper[4960]: E1201 20:58:29.328976 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:58:44 crc kubenswrapper[4960]: I1201 20:58:44.324477 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:58:44 crc kubenswrapper[4960]: E1201 20:58:44.325742 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:58:58 crc kubenswrapper[4960]: I1201 20:58:58.324758 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:58:58 crc kubenswrapper[4960]: E1201 20:58:58.325864 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:59:09 crc kubenswrapper[4960]: I1201 20:59:09.324796 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:59:09 crc kubenswrapper[4960]: E1201 20:59:09.325647 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:59:24 crc kubenswrapper[4960]: I1201 20:59:24.323919 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:59:24 crc kubenswrapper[4960]: E1201 20:59:24.324912 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:59:28 crc kubenswrapper[4960]: I1201 20:59:28.781982 4960 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2" containerName="galera" probeResult="failure" output="command timed out" Dec 01 20:59:28 crc kubenswrapper[4960]: I1201 20:59:28.785378 4960 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="c2b2a9ff-2995-4e32-bc7d-49b7fc3b86d2" containerName="galera" probeResult="failure" output="command timed out" Dec 01 20:59:35 crc kubenswrapper[4960]: I1201 20:59:35.325193 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:59:35 crc kubenswrapper[4960]: E1201 20:59:35.326525 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 20:59:50 crc kubenswrapper[4960]: I1201 20:59:50.325437 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 20:59:50 crc kubenswrapper[4960]: E1201 20:59:50.326820 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.148423 4960 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk"] Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149410 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149425 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149442 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="extract-utilities" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149449 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="extract-utilities" Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149456 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="extract-content" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149463 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="extract-content" Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149484 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="extract-utilities" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149490 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="extract-utilities" Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149504 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="extract-content" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149510 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="extract-content" Dec 01 21:00:00 crc kubenswrapper[4960]: E1201 21:00:00.149520 4960 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149527 4960 state_mem.go:107] "Deleted CPUSet assignment" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149799 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="702ffbff-4550-40d7-86fa-3fb78b676fce" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.149825 4960 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee493fec-2662-4f83-969e-fdaa2e4f2bac" containerName="registry-server" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.150881 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.153399 4960 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.154074 4960 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.158469 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk"] Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.250758 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzvg2\" (UniqueName: \"kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.250890 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.250960 4960 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.352727 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzvg2\" (UniqueName: \"kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.352892 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.353000 4960 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.353871 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.359535 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.369806 4960 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzvg2\" (UniqueName: \"kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2\") pod \"collect-profiles-29410380-r6wmk\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.487315 4960 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:00 crc kubenswrapper[4960]: W1201 21:00:00.987938 4960 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f7e830e_904c_4927_911c_9b9319d64623.slice/crio-3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b WatchSource:0}: Error finding container 3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b: Status 404 returned error can't find the container with id 3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b Dec 01 21:00:00 crc kubenswrapper[4960]: I1201 21:00:00.991044 4960 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk"] Dec 01 21:00:01 crc kubenswrapper[4960]: I1201 21:00:01.159229 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" event={"ID":"9f7e830e-904c-4927-911c-9b9319d64623","Type":"ContainerStarted","Data":"3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b"} Dec 01 21:00:02 crc kubenswrapper[4960]: I1201 21:00:02.176548 4960 generic.go:334] "Generic (PLEG): container finished" podID="9f7e830e-904c-4927-911c-9b9319d64623" containerID="57bf38ace00bb782b2ffe2848ac789ec09d583f4a851cab4a941a21c28665d1a" exitCode=0 Dec 01 21:00:02 crc kubenswrapper[4960]: I1201 21:00:02.176594 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" event={"ID":"9f7e830e-904c-4927-911c-9b9319d64623","Type":"ContainerDied","Data":"57bf38ace00bb782b2ffe2848ac789ec09d583f4a851cab4a941a21c28665d1a"} Dec 01 21:00:02 crc kubenswrapper[4960]: I1201 21:00:02.325340 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 21:00:02 crc kubenswrapper[4960]: E1201 21:00:02.325949 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.595936 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.735395 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume\") pod \"9f7e830e-904c-4927-911c-9b9319d64623\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.735668 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzvg2\" (UniqueName: \"kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2\") pod \"9f7e830e-904c-4927-911c-9b9319d64623\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.735733 4960 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume\") pod \"9f7e830e-904c-4927-911c-9b9319d64623\" (UID: \"9f7e830e-904c-4927-911c-9b9319d64623\") " Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.736317 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume" (OuterVolumeSpecName: "config-volume") pod "9f7e830e-904c-4927-911c-9b9319d64623" (UID: "9f7e830e-904c-4927-911c-9b9319d64623"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.742414 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2" (OuterVolumeSpecName: "kube-api-access-zzvg2") pod "9f7e830e-904c-4927-911c-9b9319d64623" (UID: "9f7e830e-904c-4927-911c-9b9319d64623"). InnerVolumeSpecName "kube-api-access-zzvg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.742519 4960 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9f7e830e-904c-4927-911c-9b9319d64623" (UID: "9f7e830e-904c-4927-911c-9b9319d64623"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.838368 4960 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9f7e830e-904c-4927-911c-9b9319d64623-config-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.838411 4960 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzvg2\" (UniqueName: \"kubernetes.io/projected/9f7e830e-904c-4927-911c-9b9319d64623-kube-api-access-zzvg2\") on node \"crc\" DevicePath \"\"" Dec 01 21:00:03 crc kubenswrapper[4960]: I1201 21:00:03.838421 4960 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9f7e830e-904c-4927-911c-9b9319d64623-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 01 21:00:04 crc kubenswrapper[4960]: I1201 21:00:04.200423 4960 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" event={"ID":"9f7e830e-904c-4927-911c-9b9319d64623","Type":"ContainerDied","Data":"3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b"} Dec 01 21:00:04 crc kubenswrapper[4960]: I1201 21:00:04.200473 4960 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ebd51b283590353e84dffb254397548db40283a0f9a8fefd1333352c306cb3b" Dec 01 21:00:04 crc kubenswrapper[4960]: I1201 21:00:04.200518 4960 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410380-r6wmk" Dec 01 21:00:04 crc kubenswrapper[4960]: I1201 21:00:04.677863 4960 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf"] Dec 01 21:00:04 crc kubenswrapper[4960]: I1201 21:00:04.686905 4960 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410335-4lkbf"] Dec 01 21:00:05 crc kubenswrapper[4960]: I1201 21:00:05.344456 4960 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4a047e-d49a-4f90-867e-5c04f2c572fc" path="/var/lib/kubelet/pods/9f4a047e-d49a-4f90-867e-5c04f2c572fc/volumes" Dec 01 21:00:14 crc kubenswrapper[4960]: I1201 21:00:14.323844 4960 scope.go:117] "RemoveContainer" containerID="7f7342edf7de7b5e5bc5ba025b34a343062d9724f45334566111d8f87e749f57" Dec 01 21:00:14 crc kubenswrapper[4960]: E1201 21:00:14.326278 4960 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ct7db_openshift-machine-config-operator(b6dbabf7-fd52-4f8d-9bca-093018d1c0b9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ct7db" podUID="b6dbabf7-fd52-4f8d-9bca-093018d1c0b9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113400745024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113400745017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113367050016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113367050015456 5ustar corecore